Feb 20 08:05:49 crc systemd[1]: Starting Kubernetes Kubelet... Feb 20 08:05:49 crc restorecon[4679]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 08:05:50 crc restorecon[4679]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Feb 20 08:05:51 crc kubenswrapper[4948]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 20 08:05:51 crc kubenswrapper[4948]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Feb 20 08:05:51 crc kubenswrapper[4948]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 20 08:05:51 crc kubenswrapper[4948]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 20 08:05:51 crc kubenswrapper[4948]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Feb 20 08:05:51 crc kubenswrapper[4948]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.449863 4948 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453454 4948 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453478 4948 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453485 4948 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453491 4948 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453496 4948 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453501 4948 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453508 4948 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453514 4948 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453521 4948 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453527 4948 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453533 4948 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453539 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453544 4948 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453549 4948 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453554 4948 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453560 4948 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453567 4948 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453573 4948 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453578 4948 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453583 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453588 4948 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453593 4948 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453598 4948 feature_gate.go:330] unrecognized feature gate: Example Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453603 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453609 4948 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453615 4948 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453620 4948 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453625 4948 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453631 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453636 4948 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453652 4948 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453658 4948 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453663 4948 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453669 4948 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453675 4948 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453680 4948 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453687 4948 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453692 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453697 4948 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453702 4948 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453707 4948 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453711 4948 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453716 4948 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453725 4948 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453730 4948 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453735 4948 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453740 4948 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453744 4948 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453749 4948 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453754 4948 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453759 4948 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453764 4948 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453768 4948 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453773 4948 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453778 4948 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453782 4948 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453787 4948 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453792 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453796 4948 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453801 4948 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453806 4948 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453810 4948 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453815 4948 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453820 4948 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453825 4948 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453829 4948 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453835 4948 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453840 4948 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453844 4948 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453850 4948 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453855 4948 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.453957 4948 flags.go:64] FLAG: --address="0.0.0.0" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.453988 4948 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454002 4948 flags.go:64] FLAG: --anonymous-auth="true" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454010 4948 flags.go:64] FLAG: --application-metrics-count-limit="100" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454019 4948 flags.go:64] FLAG: --authentication-token-webhook="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454024 4948 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454032 4948 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454039 4948 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454045 4948 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454051 4948 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454057 4948 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454063 4948 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454068 4948 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454074 4948 flags.go:64] FLAG: --cgroup-root="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454080 4948 flags.go:64] FLAG: --cgroups-per-qos="true" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454085 4948 flags.go:64] FLAG: --client-ca-file="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454091 4948 flags.go:64] FLAG: --cloud-config="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454096 4948 flags.go:64] FLAG: --cloud-provider="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454102 4948 flags.go:64] FLAG: --cluster-dns="[]" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454135 4948 flags.go:64] FLAG: --cluster-domain="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454140 4948 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454146 4948 flags.go:64] FLAG: --config-dir="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454151 4948 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454158 4948 flags.go:64] FLAG: --container-log-max-files="5" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454166 4948 flags.go:64] FLAG: --container-log-max-size="10Mi" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454172 4948 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454178 4948 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454220 4948 flags.go:64] FLAG: --containerd-namespace="k8s.io" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454227 4948 flags.go:64] FLAG: --contention-profiling="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454233 4948 flags.go:64] FLAG: --cpu-cfs-quota="true" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454239 4948 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454246 4948 flags.go:64] FLAG: --cpu-manager-policy="none" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454252 4948 flags.go:64] FLAG: --cpu-manager-policy-options="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454259 4948 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454265 4948 flags.go:64] FLAG: --enable-controller-attach-detach="true" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454271 4948 flags.go:64] FLAG: --enable-debugging-handlers="true" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454277 4948 flags.go:64] FLAG: --enable-load-reader="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454282 4948 flags.go:64] FLAG: --enable-server="true" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454288 4948 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454294 4948 flags.go:64] FLAG: --event-burst="100" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454300 4948 flags.go:64] FLAG: --event-qps="50" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454306 4948 flags.go:64] FLAG: --event-storage-age-limit="default=0" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454311 4948 flags.go:64] FLAG: --event-storage-event-limit="default=0" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454317 4948 flags.go:64] FLAG: --eviction-hard="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454324 4948 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454330 4948 flags.go:64] FLAG: --eviction-minimum-reclaim="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454335 4948 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454341 4948 flags.go:64] FLAG: --eviction-soft="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454347 4948 flags.go:64] FLAG: --eviction-soft-grace-period="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454354 4948 flags.go:64] FLAG: --exit-on-lock-contention="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454360 4948 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454366 4948 flags.go:64] FLAG: --experimental-mounter-path="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454371 4948 flags.go:64] FLAG: --fail-cgroupv1="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454377 4948 flags.go:64] FLAG: --fail-swap-on="true" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454382 4948 flags.go:64] FLAG: --feature-gates="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454389 4948 flags.go:64] FLAG: --file-check-frequency="20s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454395 4948 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454401 4948 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454407 4948 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454413 4948 flags.go:64] FLAG: --healthz-port="10248" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454418 4948 flags.go:64] FLAG: --help="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454424 4948 flags.go:64] FLAG: --hostname-override="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454429 4948 flags.go:64] FLAG: --housekeeping-interval="10s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454435 4948 flags.go:64] FLAG: --http-check-frequency="20s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454440 4948 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454446 4948 flags.go:64] FLAG: --image-credential-provider-config="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454451 4948 flags.go:64] FLAG: --image-gc-high-threshold="85" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454458 4948 flags.go:64] FLAG: --image-gc-low-threshold="80" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454464 4948 flags.go:64] FLAG: --image-service-endpoint="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454469 4948 flags.go:64] FLAG: --kernel-memcg-notification="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454474 4948 flags.go:64] FLAG: --kube-api-burst="100" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454480 4948 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454486 4948 flags.go:64] FLAG: --kube-api-qps="50" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454491 4948 flags.go:64] FLAG: --kube-reserved="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454497 4948 flags.go:64] FLAG: --kube-reserved-cgroup="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454502 4948 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454508 4948 flags.go:64] FLAG: --kubelet-cgroups="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454513 4948 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454519 4948 flags.go:64] FLAG: --lock-file="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454524 4948 flags.go:64] FLAG: --log-cadvisor-usage="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454530 4948 flags.go:64] FLAG: --log-flush-frequency="5s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454535 4948 flags.go:64] FLAG: --log-json-info-buffer-size="0" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454544 4948 flags.go:64] FLAG: --log-json-split-stream="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454549 4948 flags.go:64] FLAG: --log-text-info-buffer-size="0" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454555 4948 flags.go:64] FLAG: --log-text-split-stream="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454560 4948 flags.go:64] FLAG: --logging-format="text" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454566 4948 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454572 4948 flags.go:64] FLAG: --make-iptables-util-chains="true" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454577 4948 flags.go:64] FLAG: --manifest-url="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454583 4948 flags.go:64] FLAG: --manifest-url-header="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454590 4948 flags.go:64] FLAG: --max-housekeeping-interval="15s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454596 4948 flags.go:64] FLAG: --max-open-files="1000000" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454602 4948 flags.go:64] FLAG: --max-pods="110" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454608 4948 flags.go:64] FLAG: --maximum-dead-containers="-1" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454614 4948 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454620 4948 flags.go:64] FLAG: --memory-manager-policy="None" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454625 4948 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454631 4948 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454637 4948 flags.go:64] FLAG: --node-ip="192.168.126.11" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454642 4948 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454663 4948 flags.go:64] FLAG: --node-status-max-images="50" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454668 4948 flags.go:64] FLAG: --node-status-update-frequency="10s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454674 4948 flags.go:64] FLAG: --oom-score-adj="-999" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454680 4948 flags.go:64] FLAG: --pod-cidr="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454686 4948 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454694 4948 flags.go:64] FLAG: --pod-manifest-path="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454699 4948 flags.go:64] FLAG: --pod-max-pids="-1" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454706 4948 flags.go:64] FLAG: --pods-per-core="0" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454712 4948 flags.go:64] FLAG: --port="10250" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454717 4948 flags.go:64] FLAG: --protect-kernel-defaults="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454723 4948 flags.go:64] FLAG: --provider-id="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454728 4948 flags.go:64] FLAG: --qos-reserved="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454734 4948 flags.go:64] FLAG: --read-only-port="10255" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454740 4948 flags.go:64] FLAG: --register-node="true" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454745 4948 flags.go:64] FLAG: --register-schedulable="true" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454750 4948 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454760 4948 flags.go:64] FLAG: --registry-burst="10" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454765 4948 flags.go:64] FLAG: --registry-qps="5" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454771 4948 flags.go:64] FLAG: --reserved-cpus="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454776 4948 flags.go:64] FLAG: --reserved-memory="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454783 4948 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454789 4948 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454795 4948 flags.go:64] FLAG: --rotate-certificates="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454800 4948 flags.go:64] FLAG: --rotate-server-certificates="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454806 4948 flags.go:64] FLAG: --runonce="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454811 4948 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454817 4948 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454822 4948 flags.go:64] FLAG: --seccomp-default="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454828 4948 flags.go:64] FLAG: --serialize-image-pulls="true" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454834 4948 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454839 4948 flags.go:64] FLAG: --storage-driver-db="cadvisor" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454845 4948 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454850 4948 flags.go:64] FLAG: --storage-driver-password="root" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454858 4948 flags.go:64] FLAG: --storage-driver-secure="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454865 4948 flags.go:64] FLAG: --storage-driver-table="stats" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454870 4948 flags.go:64] FLAG: --storage-driver-user="root" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454875 4948 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454882 4948 flags.go:64] FLAG: --sync-frequency="1m0s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454887 4948 flags.go:64] FLAG: --system-cgroups="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454894 4948 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454903 4948 flags.go:64] FLAG: --system-reserved-cgroup="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454908 4948 flags.go:64] FLAG: --tls-cert-file="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454913 4948 flags.go:64] FLAG: --tls-cipher-suites="[]" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454920 4948 flags.go:64] FLAG: --tls-min-version="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454926 4948 flags.go:64] FLAG: --tls-private-key-file="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454932 4948 flags.go:64] FLAG: --topology-manager-policy="none" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454937 4948 flags.go:64] FLAG: --topology-manager-policy-options="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454943 4948 flags.go:64] FLAG: --topology-manager-scope="container" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454948 4948 flags.go:64] FLAG: --v="2" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454956 4948 flags.go:64] FLAG: --version="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454963 4948 flags.go:64] FLAG: --vmodule="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454986 4948 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454993 4948 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455157 4948 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455164 4948 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455170 4948 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455176 4948 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455182 4948 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455187 4948 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455195 4948 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455200 4948 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455205 4948 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455210 4948 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455215 4948 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455219 4948 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455227 4948 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455231 4948 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455236 4948 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455241 4948 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455246 4948 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455251 4948 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455256 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455260 4948 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455265 4948 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455270 4948 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455276 4948 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455281 4948 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455286 4948 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455290 4948 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455295 4948 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455300 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455305 4948 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455310 4948 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455315 4948 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455319 4948 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455324 4948 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455329 4948 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455334 4948 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455340 4948 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455347 4948 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455353 4948 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455360 4948 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455366 4948 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455371 4948 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455376 4948 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455381 4948 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455386 4948 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455393 4948 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455397 4948 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455403 4948 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455409 4948 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455415 4948 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455420 4948 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455426 4948 feature_gate.go:330] unrecognized feature gate: Example Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455432 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455437 4948 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455442 4948 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455446 4948 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455451 4948 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455456 4948 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455460 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455466 4948 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455471 4948 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455477 4948 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455483 4948 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455490 4948 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455496 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455502 4948 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455507 4948 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455513 4948 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455518 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455523 4948 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455528 4948 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455534 4948 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.456259 4948 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.468425 4948 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.468883 4948 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469037 4948 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469053 4948 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469063 4948 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469072 4948 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469082 4948 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469090 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469098 4948 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469106 4948 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469113 4948 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469124 4948 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469138 4948 feature_gate.go:330] unrecognized feature gate: Example Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469148 4948 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469157 4948 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469166 4948 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469175 4948 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469183 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469191 4948 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469201 4948 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469209 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469216 4948 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469224 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469232 4948 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469240 4948 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469248 4948 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469257 4948 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469267 4948 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469277 4948 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469287 4948 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469296 4948 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469307 4948 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469318 4948 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469327 4948 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469335 4948 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469344 4948 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469354 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469362 4948 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469371 4948 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469379 4948 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469386 4948 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469394 4948 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469402 4948 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469411 4948 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469419 4948 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469427 4948 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469435 4948 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469443 4948 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469451 4948 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469459 4948 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469466 4948 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469474 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469482 4948 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469490 4948 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469498 4948 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469506 4948 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469515 4948 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469523 4948 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469531 4948 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469539 4948 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469547 4948 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469554 4948 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469562 4948 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469570 4948 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469577 4948 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469585 4948 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469592 4948 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469600 4948 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469608 4948 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469616 4948 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469626 4948 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469636 4948 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469646 4948 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.469660 4948 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469879 4948 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469893 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469902 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469911 4948 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469920 4948 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469929 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469936 4948 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469944 4948 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469952 4948 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469959 4948 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469967 4948 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469997 4948 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470005 4948 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470013 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470021 4948 feature_gate.go:330] unrecognized feature gate: Example Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470029 4948 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470041 4948 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470048 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470056 4948 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470064 4948 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470072 4948 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470079 4948 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470087 4948 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470095 4948 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470102 4948 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470110 4948 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470118 4948 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470125 4948 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470133 4948 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470141 4948 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470151 4948 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470162 4948 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470170 4948 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470178 4948 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470187 4948 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470195 4948 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470205 4948 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470215 4948 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470225 4948 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470233 4948 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470243 4948 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470252 4948 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470260 4948 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470268 4948 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470276 4948 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470284 4948 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470292 4948 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470300 4948 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470308 4948 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470316 4948 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470323 4948 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470331 4948 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470339 4948 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470346 4948 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470354 4948 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470362 4948 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470369 4948 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470380 4948 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470388 4948 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470398 4948 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470406 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470414 4948 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470422 4948 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470431 4948 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470439 4948 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470447 4948 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470455 4948 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470463 4948 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470470 4948 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470477 4948 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470486 4948 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.470498 4948 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.471560 4948 server.go:940] "Client rotation is on, will bootstrap in background" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.483304 4948 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.483507 4948 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.485464 4948 server.go:997] "Starting client certificate rotation" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.485531 4948 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.485821 4948 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-15 14:17:16.878034339 +0000 UTC Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.486006 4948 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.511641 4948 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 20 08:05:51 crc kubenswrapper[4948]: E0220 08:05:51.516409 4948 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.113:6443: connect: connection refused" logger="UnhandledError" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.518438 4948 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.542550 4948 log.go:25] "Validated CRI v1 runtime API" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.583288 4948 log.go:25] "Validated CRI v1 image API" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.585935 4948 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.591651 4948 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-02-20-08-01-26-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.591698 4948 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.619030 4948 manager.go:217] Machine: {Timestamp:2026-02-20 08:05:51.615328761 +0000 UTC m=+0.589823641 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654132736 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:51d89745-b91f-4e85-9724-1ef53e3862a9 BootID:8f36346c-e92f-4a00-a0af-e0652f71277d Filesystems:[{Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730829824 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827068416 Type:vfs Inodes:1048576 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:85:61:b1 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:85:61:b1 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:e8:40:45 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:13:c9:98 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:bf:26:78 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:f4:2b:94 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:7e:8d:3c:94:36:7d Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:8e:8f:b6:78:f7:89 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654132736 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.619446 4948 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.619680 4948 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.623355 4948 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.623744 4948 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.623856 4948 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.625332 4948 topology_manager.go:138] "Creating topology manager with none policy" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.625370 4948 container_manager_linux.go:303] "Creating device plugin manager" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.626036 4948 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.626113 4948 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.626705 4948 state_mem.go:36] "Initialized new in-memory state store" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.626963 4948 server.go:1245] "Using root directory" path="/var/lib/kubelet" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.631006 4948 kubelet.go:418] "Attempting to sync node with API server" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.631089 4948 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.631186 4948 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.631212 4948 kubelet.go:324] "Adding apiserver pod source" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.631233 4948 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.636245 4948 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.637543 4948 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.637890 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.637941 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Feb 20 08:05:51 crc kubenswrapper[4948]: E0220 08:05:51.638084 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.113:6443: connect: connection refused" logger="UnhandledError" Feb 20 08:05:51 crc kubenswrapper[4948]: E0220 08:05:51.638084 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.113:6443: connect: connection refused" logger="UnhandledError" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.640460 4948 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.642424 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.642467 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.642483 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.642496 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.642518 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.642531 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.642545 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.642568 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.642585 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.642600 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.642622 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.642635 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.643889 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.644615 4948 server.go:1280] "Started kubelet" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.646238 4948 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.646284 4948 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.646519 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.647071 4948 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 20 08:05:51 crc systemd[1]: Started Kubernetes Kubelet. Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.648924 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.649336 4948 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.649375 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 06:37:54.603984492 +0000 UTC Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.649565 4948 volume_manager.go:287] "The desired_state_of_world populator starts" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.649586 4948 volume_manager.go:289] "Starting Kubelet Volume Manager" Feb 20 08:05:51 crc kubenswrapper[4948]: E0220 08:05:51.649639 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.649682 4948 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Feb 20 08:05:51 crc kubenswrapper[4948]: E0220 08:05:51.650826 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" interval="200ms" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.651391 4948 server.go:460] "Adding debug handlers to kubelet server" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.651464 4948 factory.go:55] Registering systemd factory Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.651501 4948 factory.go:221] Registration of the systemd container factory successfully Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.652075 4948 factory.go:153] Registering CRI-O factory Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.652103 4948 factory.go:221] Registration of the crio container factory successfully Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.652234 4948 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.652276 4948 factory.go:103] Registering Raw factory Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.652303 4948 manager.go:1196] Started watching for new ooms in manager Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.652344 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Feb 20 08:05:51 crc kubenswrapper[4948]: E0220 08:05:51.652488 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.113:6443: connect: connection refused" logger="UnhandledError" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.653403 4948 manager.go:319] Starting recovery of all containers Feb 20 08:05:51 crc kubenswrapper[4948]: E0220 08:05:51.652828 4948 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.113:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.1895e5d9b4d34f97 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-20 08:05:51.644569495 +0000 UTC m=+0.619064345,LastTimestamp:2026-02-20 08:05:51.644569495 +0000 UTC m=+0.619064345,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676451 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676519 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676544 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676563 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676583 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676603 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676622 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676641 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676663 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676683 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676751 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676772 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676791 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676834 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676854 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676874 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676903 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676923 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676940 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676960 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677014 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677041 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677064 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677132 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677153 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677172 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677195 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677217 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677235 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677257 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677275 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677296 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677316 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677335 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677355 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677374 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677392 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677447 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677465 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677483 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677501 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677518 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677537 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677554 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677577 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677596 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677616 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677637 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677656 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677678 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677699 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677718 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677743 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677766 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.679945 4948 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680053 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680087 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680109 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680132 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680153 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680177 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680199 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680218 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680240 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680259 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680278 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680297 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680316 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680335 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680354 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680372 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680392 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680412 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680432 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680471 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680491 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680510 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680529 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680545 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680569 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680589 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680609 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680627 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680646 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680666 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680685 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680703 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680722 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680740 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680757 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680775 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680793 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680811 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680829 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680849 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680872 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680894 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680914 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680933 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680954 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681011 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681037 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681058 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681078 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681134 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681163 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681184 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681205 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681229 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681250 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681272 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681294 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681315 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681335 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681357 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681377 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681395 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681414 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681432 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681451 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681467 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681485 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681504 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681522 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681540 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681563 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681580 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681598 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681617 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681637 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681656 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681675 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681697 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681714 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681731 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681749 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681768 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681787 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681804 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681822 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681840 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681865 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681882 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681901 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681922 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681940 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681958 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682010 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682034 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682052 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682070 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682088 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682105 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682124 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682142 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682159 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682177 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682195 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682212 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682229 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682247 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682263 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682280 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682297 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682315 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682335 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682353 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682371 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682389 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682405 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682425 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682444 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682461 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682484 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682502 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682535 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682555 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682574 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682591 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682610 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682629 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682648 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682668 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682696 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682714 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682732 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682749 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682766 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682785 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682803 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682821 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682841 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682859 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682877 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682897 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682915 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682934 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682952 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682996 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.683055 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.683078 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.683097 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.683113 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.683134 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.683152 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.683177 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.683194 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.683215 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.683233 4948 reconstruct.go:97] "Volume reconstruction finished" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.683246 4948 reconciler.go:26] "Reconciler: start to sync state" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.692575 4948 manager.go:324] Recovery completed Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.711788 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.713964 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.714172 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.714269 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.715469 4948 cpu_manager.go:225] "Starting CPU manager" policy="none" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.715598 4948 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.715683 4948 state_mem.go:36] "Initialized new in-memory state store" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.717181 4948 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.721053 4948 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.721146 4948 status_manager.go:217] "Starting to sync pod status with apiserver" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.721216 4948 kubelet.go:2335] "Starting kubelet main sync loop" Feb 20 08:05:51 crc kubenswrapper[4948]: E0220 08:05:51.721308 4948 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.722199 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Feb 20 08:05:51 crc kubenswrapper[4948]: E0220 08:05:51.722302 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.113:6443: connect: connection refused" logger="UnhandledError" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.739348 4948 policy_none.go:49] "None policy: Start" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.741526 4948 memory_manager.go:170] "Starting memorymanager" policy="None" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.741664 4948 state_mem.go:35] "Initializing new in-memory state store" Feb 20 08:05:51 crc kubenswrapper[4948]: E0220 08:05:51.749762 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.811583 4948 manager.go:334] "Starting Device Plugin manager" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.811937 4948 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.811968 4948 server.go:79] "Starting device plugin registration server" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.812582 4948 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.812612 4948 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.813587 4948 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.813790 4948 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.813807 4948 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.821802 4948 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.821950 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.824489 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.824572 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.824597 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.824880 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.825249 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.825322 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.830784 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.830788 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:51 crc kubenswrapper[4948]: E0220 08:05:51.830827 4948 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.830857 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.830870 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.830890 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.830893 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.831966 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.832265 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.832361 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.834269 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.834314 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.834268 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.834364 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.834383 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.834334 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.834639 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.834879 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.835007 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.835904 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.835945 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.836005 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.836155 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.836279 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.836318 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.836336 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.836756 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.836813 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.837581 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.837632 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.837653 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.837850 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.837885 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.838645 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.838685 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.838704 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.839395 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.839451 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.839473 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:51 crc kubenswrapper[4948]: E0220 08:05:51.851707 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" interval="400ms" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.885481 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.886009 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.886448 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.886538 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.886729 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.887095 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.887325 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.887690 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.887907 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.888270 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.888487 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.888674 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.888851 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.888961 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.889043 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.913359 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.915026 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.915095 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.915117 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.915164 4948 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 20 08:05:51 crc kubenswrapper[4948]: E0220 08:05:51.916203 4948 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.113:6443: connect: connection refused" node="crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.990693 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.990755 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.990804 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.990840 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.990912 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.990944 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991007 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991006 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991106 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991100 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991104 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991038 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991216 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991270 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991016 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991206 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991243 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991259 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991370 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991408 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991439 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991469 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991485 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991518 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991526 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991501 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991590 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991553 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991655 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991555 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.116635 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.119124 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.119190 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.119211 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.119255 4948 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 20 08:05:52 crc kubenswrapper[4948]: E0220 08:05:52.119769 4948 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.113:6443: connect: connection refused" node="crc" Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.179856 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.195088 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.211650 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.233038 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 08:05:52 crc kubenswrapper[4948]: W0220 08:05:52.238014 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-fb97570385558360c8e37ed85db7ebad3ca89392dbdd9304506115ac70cfcc6a WatchSource:0}: Error finding container fb97570385558360c8e37ed85db7ebad3ca89392dbdd9304506115ac70cfcc6a: Status 404 returned error can't find the container with id fb97570385558360c8e37ed85db7ebad3ca89392dbdd9304506115ac70cfcc6a Feb 20 08:05:52 crc kubenswrapper[4948]: W0220 08:05:52.240089 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-7dd21173ef20d7b33a38877aa593fa065ed43716bf18e03a37a95cf27c3b5b2d WatchSource:0}: Error finding container 7dd21173ef20d7b33a38877aa593fa065ed43716bf18e03a37a95cf27c3b5b2d: Status 404 returned error can't find the container with id 7dd21173ef20d7b33a38877aa593fa065ed43716bf18e03a37a95cf27c3b5b2d Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.242620 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 20 08:05:52 crc kubenswrapper[4948]: E0220 08:05:52.252675 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" interval="800ms" Feb 20 08:05:52 crc kubenswrapper[4948]: W0220 08:05:52.254516 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-e5cb1d98ca7c049a62301d2cdb2cadea4ba1ba0b36b390cd5329fb1b233ddfb4 WatchSource:0}: Error finding container e5cb1d98ca7c049a62301d2cdb2cadea4ba1ba0b36b390cd5329fb1b233ddfb4: Status 404 returned error can't find the container with id e5cb1d98ca7c049a62301d2cdb2cadea4ba1ba0b36b390cd5329fb1b233ddfb4 Feb 20 08:05:52 crc kubenswrapper[4948]: W0220 08:05:52.261089 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-efec26f7c799b90d098d9b607041426a10d2dfaa0c9145be7330d16b271ea176 WatchSource:0}: Error finding container efec26f7c799b90d098d9b607041426a10d2dfaa0c9145be7330d16b271ea176: Status 404 returned error can't find the container with id efec26f7c799b90d098d9b607041426a10d2dfaa0c9145be7330d16b271ea176 Feb 20 08:05:52 crc kubenswrapper[4948]: W0220 08:05:52.263054 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-090a66e9126788a7d8e207e7d0a743e706639096359b71eab9813e032dffe803 WatchSource:0}: Error finding container 090a66e9126788a7d8e207e7d0a743e706639096359b71eab9813e032dffe803: Status 404 returned error can't find the container with id 090a66e9126788a7d8e207e7d0a743e706639096359b71eab9813e032dffe803 Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.520585 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.521875 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.521938 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.521957 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.522032 4948 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 20 08:05:52 crc kubenswrapper[4948]: E0220 08:05:52.522714 4948 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.113:6443: connect: connection refused" node="crc" Feb 20 08:05:52 crc kubenswrapper[4948]: W0220 08:05:52.592119 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Feb 20 08:05:52 crc kubenswrapper[4948]: E0220 08:05:52.592270 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.113:6443: connect: connection refused" logger="UnhandledError" Feb 20 08:05:52 crc kubenswrapper[4948]: W0220 08:05:52.612332 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Feb 20 08:05:52 crc kubenswrapper[4948]: E0220 08:05:52.612443 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.113:6443: connect: connection refused" logger="UnhandledError" Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.648273 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.650481 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 14:19:00.424502236 +0000 UTC Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.728635 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"090a66e9126788a7d8e207e7d0a743e706639096359b71eab9813e032dffe803"} Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.730468 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"efec26f7c799b90d098d9b607041426a10d2dfaa0c9145be7330d16b271ea176"} Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.731611 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e5cb1d98ca7c049a62301d2cdb2cadea4ba1ba0b36b390cd5329fb1b233ddfb4"} Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.733156 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7dd21173ef20d7b33a38877aa593fa065ed43716bf18e03a37a95cf27c3b5b2d"} Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.734501 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fb97570385558360c8e37ed85db7ebad3ca89392dbdd9304506115ac70cfcc6a"} Feb 20 08:05:52 crc kubenswrapper[4948]: W0220 08:05:52.941380 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Feb 20 08:05:52 crc kubenswrapper[4948]: E0220 08:05:52.941502 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.113:6443: connect: connection refused" logger="UnhandledError" Feb 20 08:05:53 crc kubenswrapper[4948]: E0220 08:05:53.054386 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" interval="1.6s" Feb 20 08:05:53 crc kubenswrapper[4948]: W0220 08:05:53.119819 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Feb 20 08:05:53 crc kubenswrapper[4948]: E0220 08:05:53.119963 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.113:6443: connect: connection refused" logger="UnhandledError" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.323402 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.325018 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.325054 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.325064 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.325087 4948 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 20 08:05:53 crc kubenswrapper[4948]: E0220 08:05:53.325615 4948 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.113:6443: connect: connection refused" node="crc" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.593156 4948 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 20 08:05:53 crc kubenswrapper[4948]: E0220 08:05:53.595038 4948 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.113:6443: connect: connection refused" logger="UnhandledError" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.647628 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.650723 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 20:09:05.356180009 +0000 UTC Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.743323 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378"} Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.743395 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737"} Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.743418 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4"} Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.745533 4948 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b" exitCode=0 Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.745699 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.745954 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b"} Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.747208 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.747258 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.747277 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.748547 4948 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe" exitCode=0 Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.748614 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe"} Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.748739 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.750232 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.750412 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.750459 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.750474 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.751408 4948 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221" exitCode=0 Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.751480 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221"} Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.751583 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.752739 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.752784 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.752802 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.753204 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.753279 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.753307 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.754758 4948 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054" exitCode=0 Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.754828 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054"} Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.755049 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.756289 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.756315 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.756344 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:54 crc kubenswrapper[4948]: W0220 08:05:54.506076 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Feb 20 08:05:54 crc kubenswrapper[4948]: E0220 08:05:54.506611 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.113:6443: connect: connection refused" logger="UnhandledError" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.647486 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.650904 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 22:14:02.857179738 +0000 UTC Feb 20 08:05:54 crc kubenswrapper[4948]: E0220 08:05:54.655493 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" interval="3.2s" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.767831 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"e67b60ad2efa00639a153f39923bfdba3b40b6b3f5126207b4f0e05abdf9f43f"} Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.767987 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.769341 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.769372 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.769381 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.774015 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.774707 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"0a52eab83c2f040b6bd090f63f55a7da942f1b9a219ae29f14fc2b9930b07da8"} Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.774764 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b5f1450fc3f715e0082ad316af3f03ca0d2d8f32b5efa9850e0694864f79153e"} Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.774784 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"001bf218b5793a1101d9960b6d81173d67c5f6259c21d780623dad9808adfd6f"} Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.776650 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.776696 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.776726 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.779671 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad"} Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.779739 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.780892 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.780923 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.780934 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.783648 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244"} Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.783675 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556"} Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.783687 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a"} Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.783697 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85"} Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.793230 4948 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc" exitCode=0 Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.793304 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc"} Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.793402 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.794737 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.794771 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.794789 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:54 crc kubenswrapper[4948]: E0220 08:05:54.874745 4948 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.113:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.1895e5d9b4d34f97 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-20 08:05:51.644569495 +0000 UTC m=+0.619064345,LastTimestamp:2026-02-20 08:05:51.644569495 +0000 UTC m=+0.619064345,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.926041 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.927284 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.927324 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.927334 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.927362 4948 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 20 08:05:54 crc kubenswrapper[4948]: E0220 08:05:54.927943 4948 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.113:6443: connect: connection refused" node="crc" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.970038 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.979744 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:05:55 crc kubenswrapper[4948]: W0220 08:05:55.015425 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Feb 20 08:05:55 crc kubenswrapper[4948]: E0220 08:05:55.015515 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.113:6443: connect: connection refused" logger="UnhandledError" Feb 20 08:05:55 crc kubenswrapper[4948]: W0220 08:05:55.058399 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Feb 20 08:05:55 crc kubenswrapper[4948]: E0220 08:05:55.058512 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.113:6443: connect: connection refused" logger="UnhandledError" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.329238 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.651806 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 14:49:40.557334232 +0000 UTC Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.799625 4948 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb" exitCode=0 Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.799783 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb"} Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.799808 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.801331 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.801380 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.801402 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.804256 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678"} Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.804340 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.804382 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.804393 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.806476 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.811385 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.811423 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.811432 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.811431 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.811466 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.811468 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.811507 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.811479 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.811538 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.811569 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.811588 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.811525 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.653075 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 03:35:14.409499108 +0000 UTC Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.812854 4948 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.812910 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.812927 4948 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.813041 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.813470 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.813722 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e"} Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.813798 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7"} Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.813820 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154"} Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.814208 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.814256 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.814272 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.815153 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.815213 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.815232 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.815479 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.815554 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.815577 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:57 crc kubenswrapper[4948]: I0220 08:05:57.295206 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:05:57 crc kubenswrapper[4948]: I0220 08:05:57.340632 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:05:57 crc kubenswrapper[4948]: I0220 08:05:57.654138 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 12:28:10.201801422 +0000 UTC Feb 20 08:05:57 crc kubenswrapper[4948]: I0220 08:05:57.701359 4948 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 20 08:05:57 crc kubenswrapper[4948]: I0220 08:05:57.822648 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5"} Feb 20 08:05:57 crc kubenswrapper[4948]: I0220 08:05:57.822740 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc"} Feb 20 08:05:57 crc kubenswrapper[4948]: I0220 08:05:57.822681 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:57 crc kubenswrapper[4948]: I0220 08:05:57.822682 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:57 crc kubenswrapper[4948]: I0220 08:05:57.824413 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:57 crc kubenswrapper[4948]: I0220 08:05:57.824484 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:57 crc kubenswrapper[4948]: I0220 08:05:57.824491 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:57 crc kubenswrapper[4948]: I0220 08:05:57.824598 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:57 crc kubenswrapper[4948]: I0220 08:05:57.824534 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:57 crc kubenswrapper[4948]: I0220 08:05:57.824655 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:58 crc kubenswrapper[4948]: I0220 08:05:58.128557 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:58 crc kubenswrapper[4948]: I0220 08:05:58.130131 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:58 crc kubenswrapper[4948]: I0220 08:05:58.130199 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:58 crc kubenswrapper[4948]: I0220 08:05:58.130222 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:58 crc kubenswrapper[4948]: I0220 08:05:58.130266 4948 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 20 08:05:58 crc kubenswrapper[4948]: I0220 08:05:58.307940 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Feb 20 08:05:58 crc kubenswrapper[4948]: I0220 08:05:58.654781 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 17:20:54.709280079 +0000 UTC Feb 20 08:05:58 crc kubenswrapper[4948]: I0220 08:05:58.825923 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:58 crc kubenswrapper[4948]: I0220 08:05:58.825923 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:58 crc kubenswrapper[4948]: I0220 08:05:58.827204 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:58 crc kubenswrapper[4948]: I0220 08:05:58.827791 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:58 crc kubenswrapper[4948]: I0220 08:05:58.828040 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:58 crc kubenswrapper[4948]: I0220 08:05:58.828104 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:58 crc kubenswrapper[4948]: I0220 08:05:58.828054 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:58 crc kubenswrapper[4948]: I0220 08:05:58.828206 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:59 crc kubenswrapper[4948]: I0220 08:05:59.655247 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 06:23:03.129009847 +0000 UTC Feb 20 08:05:59 crc kubenswrapper[4948]: I0220 08:05:59.829307 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:59 crc kubenswrapper[4948]: I0220 08:05:59.830641 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:59 crc kubenswrapper[4948]: I0220 08:05:59.830711 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:59 crc kubenswrapper[4948]: I0220 08:05:59.830737 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:00 crc kubenswrapper[4948]: I0220 08:06:00.245285 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:06:00 crc kubenswrapper[4948]: I0220 08:06:00.245525 4948 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 20 08:06:00 crc kubenswrapper[4948]: I0220 08:06:00.245595 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:06:00 crc kubenswrapper[4948]: I0220 08:06:00.247433 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:00 crc kubenswrapper[4948]: I0220 08:06:00.247635 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:00 crc kubenswrapper[4948]: I0220 08:06:00.247757 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:00 crc kubenswrapper[4948]: I0220 08:06:00.498323 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:06:00 crc kubenswrapper[4948]: I0220 08:06:00.498642 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:06:00 crc kubenswrapper[4948]: I0220 08:06:00.500505 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:00 crc kubenswrapper[4948]: I0220 08:06:00.500571 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:00 crc kubenswrapper[4948]: I0220 08:06:00.500592 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:00 crc kubenswrapper[4948]: I0220 08:06:00.656481 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 05:08:56.78082992 +0000 UTC Feb 20 08:06:01 crc kubenswrapper[4948]: I0220 08:06:01.254390 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:06:01 crc kubenswrapper[4948]: I0220 08:06:01.254647 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:06:01 crc kubenswrapper[4948]: I0220 08:06:01.256569 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:01 crc kubenswrapper[4948]: I0220 08:06:01.256627 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:01 crc kubenswrapper[4948]: I0220 08:06:01.256647 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:01 crc kubenswrapper[4948]: I0220 08:06:01.657010 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 16:35:00.138202598 +0000 UTC Feb 20 08:06:01 crc kubenswrapper[4948]: E0220 08:06:01.831406 4948 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 20 08:06:01 crc kubenswrapper[4948]: I0220 08:06:01.985063 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Feb 20 08:06:01 crc kubenswrapper[4948]: I0220 08:06:01.985346 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:06:01 crc kubenswrapper[4948]: I0220 08:06:01.987071 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:01 crc kubenswrapper[4948]: I0220 08:06:01.987136 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:01 crc kubenswrapper[4948]: I0220 08:06:01.987159 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:02 crc kubenswrapper[4948]: I0220 08:06:02.657358 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 00:07:42.351053568 +0000 UTC Feb 20 08:06:03 crc kubenswrapper[4948]: I0220 08:06:03.658469 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 00:48:14.890162779 +0000 UTC Feb 20 08:06:03 crc kubenswrapper[4948]: I0220 08:06:03.838658 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:06:03 crc kubenswrapper[4948]: I0220 08:06:03.839013 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:06:03 crc kubenswrapper[4948]: I0220 08:06:03.841290 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:03 crc kubenswrapper[4948]: I0220 08:06:03.841360 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:03 crc kubenswrapper[4948]: I0220 08:06:03.841377 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:03 crc kubenswrapper[4948]: I0220 08:06:03.847694 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:06:03 crc kubenswrapper[4948]: I0220 08:06:03.866898 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:06:03 crc kubenswrapper[4948]: I0220 08:06:03.868420 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:03 crc kubenswrapper[4948]: I0220 08:06:03.868467 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:03 crc kubenswrapper[4948]: I0220 08:06:03.868486 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:04 crc kubenswrapper[4948]: I0220 08:06:04.658700 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 08:29:25.256040411 +0000 UTC Feb 20 08:06:05 crc kubenswrapper[4948]: W0220 08:06:05.321899 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Feb 20 08:06:05 crc kubenswrapper[4948]: I0220 08:06:05.322109 4948 trace.go:236] Trace[904507372]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (20-Feb-2026 08:05:55.320) (total time: 10001ms): Feb 20 08:06:05 crc kubenswrapper[4948]: Trace[904507372]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (08:06:05.321) Feb 20 08:06:05 crc kubenswrapper[4948]: Trace[904507372]: [10.001802632s] [10.001802632s] END Feb 20 08:06:05 crc kubenswrapper[4948]: E0220 08:06:05.322150 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Feb 20 08:06:05 crc kubenswrapper[4948]: I0220 08:06:05.648844 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Feb 20 08:06:05 crc kubenswrapper[4948]: I0220 08:06:05.659051 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 12:45:04.669657577 +0000 UTC Feb 20 08:06:06 crc kubenswrapper[4948]: I0220 08:06:06.659652 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 21:10:33.544935053 +0000 UTC Feb 20 08:06:06 crc kubenswrapper[4948]: I0220 08:06:06.839386 4948 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 20 08:06:06 crc kubenswrapper[4948]: I0220 08:06:06.839505 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 20 08:06:06 crc kubenswrapper[4948]: I0220 08:06:06.976350 4948 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 20 08:06:06 crc kubenswrapper[4948]: I0220 08:06:06.976472 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 20 08:06:06 crc kubenswrapper[4948]: I0220 08:06:06.984816 4948 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Feb 20 08:06:06 crc kubenswrapper[4948]: I0220 08:06:06.984873 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 20 08:06:07 crc kubenswrapper[4948]: I0220 08:06:07.350501 4948 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]log ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]etcd ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/openshift.io-api-request-count-filter ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/openshift.io-startkubeinformers ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/start-apiserver-admission-initializer ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/generic-apiserver-start-informers ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/priority-and-fairness-config-consumer ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/priority-and-fairness-filter ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/storage-object-count-tracker-hook ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/start-apiextensions-informers ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/start-apiextensions-controllers ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/crd-informer-synced ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/start-system-namespaces-controller ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/start-cluster-authentication-info-controller ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/start-legacy-token-tracking-controller ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/start-service-ip-repair-controllers ok Feb 20 08:06:07 crc kubenswrapper[4948]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Feb 20 08:06:07 crc kubenswrapper[4948]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/priority-and-fairness-config-producer ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/bootstrap-controller ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/start-kube-aggregator-informers ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/apiservice-status-local-available-controller ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/apiservice-status-remote-available-controller ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/apiservice-registration-controller ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/apiservice-wait-for-first-sync ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/apiservice-discovery-controller ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/kube-apiserver-autoregistration ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]autoregister-completion ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/apiservice-openapi-controller ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/apiservice-openapiv3-controller ok Feb 20 08:06:07 crc kubenswrapper[4948]: livez check failed Feb 20 08:06:07 crc kubenswrapper[4948]: I0220 08:06:07.350601 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:06:07 crc kubenswrapper[4948]: I0220 08:06:07.660085 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 02:32:07.650627102 +0000 UTC Feb 20 08:06:08 crc kubenswrapper[4948]: I0220 08:06:08.660480 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 13:51:37.626336664 +0000 UTC Feb 20 08:06:09 crc kubenswrapper[4948]: I0220 08:06:09.661592 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 22:08:24.341362603 +0000 UTC Feb 20 08:06:10 crc kubenswrapper[4948]: I0220 08:06:10.662235 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 09:32:00.608275077 +0000 UTC Feb 20 08:06:11 crc kubenswrapper[4948]: I0220 08:06:11.087836 4948 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 20 08:06:11 crc kubenswrapper[4948]: I0220 08:06:11.662456 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 22:21:46.593300688 +0000 UTC Feb 20 08:06:11 crc kubenswrapper[4948]: E0220 08:06:11.831708 4948 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 20 08:06:11 crc kubenswrapper[4948]: E0220 08:06:11.980201 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Feb 20 08:06:11 crc kubenswrapper[4948]: I0220 08:06:11.985499 4948 trace.go:236] Trace[1559600258]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (20-Feb-2026 08:05:58.510) (total time: 13475ms): Feb 20 08:06:11 crc kubenswrapper[4948]: Trace[1559600258]: ---"Objects listed" error: 13475ms (08:06:11.985) Feb 20 08:06:11 crc kubenswrapper[4948]: Trace[1559600258]: [13.475421278s] [13.475421278s] END Feb 20 08:06:11 crc kubenswrapper[4948]: I0220 08:06:11.985544 4948 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 20 08:06:11 crc kubenswrapper[4948]: E0220 08:06:11.988019 4948 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Feb 20 08:06:11 crc kubenswrapper[4948]: I0220 08:06:11.988380 4948 trace.go:236] Trace[423861376]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (20-Feb-2026 08:05:59.591) (total time: 12397ms): Feb 20 08:06:11 crc kubenswrapper[4948]: Trace[423861376]: ---"Objects listed" error: 12397ms (08:06:11.988) Feb 20 08:06:11 crc kubenswrapper[4948]: Trace[423861376]: [12.397093228s] [12.397093228s] END Feb 20 08:06:11 crc kubenswrapper[4948]: I0220 08:06:11.988404 4948 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 20 08:06:11 crc kubenswrapper[4948]: I0220 08:06:11.988401 4948 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Feb 20 08:06:11 crc kubenswrapper[4948]: I0220 08:06:11.990238 4948 trace.go:236] Trace[877424968]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (20-Feb-2026 08:05:59.335) (total time: 12654ms): Feb 20 08:06:11 crc kubenswrapper[4948]: Trace[877424968]: ---"Objects listed" error: 12654ms (08:06:11.990) Feb 20 08:06:11 crc kubenswrapper[4948]: Trace[877424968]: [12.654435359s] [12.654435359s] END Feb 20 08:06:11 crc kubenswrapper[4948]: I0220 08:06:11.990261 4948 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 20 08:06:11 crc kubenswrapper[4948]: I0220 08:06:11.991914 4948 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.021130 4948 csr.go:261] certificate signing request csr-s6842 is approved, waiting to be issued Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.032521 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.035907 4948 csr.go:257] certificate signing request csr-s6842 is issued Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.041562 4948 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:42940->192.168.126.11:17697: read: connection reset by peer" start-of-body= Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.041652 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:42940->192.168.126.11:17697: read: connection reset by peer" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.041565 4948 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:42930->192.168.126.11:17697: read: connection reset by peer" start-of-body= Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.041797 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:42930->192.168.126.11:17697: read: connection reset by peer" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.072810 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.349345 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.350553 4948 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.350678 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.356507 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.646914 4948 apiserver.go:52] "Watching apiserver" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.650419 4948 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.650723 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c"] Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.651066 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.651190 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.651366 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.651500 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.651513 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.651557 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.651607 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.651796 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.651894 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.653200 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.653222 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.654002 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.654134 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.654641 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.654654 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.655039 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.656245 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.656474 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.663432 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 11:42:50.465544472 +0000 UTC Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.682227 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.694163 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.711057 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.726118 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.739686 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.750551 4948 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.751906 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.773743 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795008 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795056 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795074 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795094 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795112 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795130 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795149 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795167 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795189 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795212 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795234 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795254 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795271 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795287 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795302 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795319 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795335 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795350 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795405 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795424 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795443 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795458 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795474 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795492 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795509 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795530 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795552 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795578 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795601 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795623 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795641 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795921 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.796140 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.796413 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.796458 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.796465 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.796498 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.796539 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.796690 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.796738 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.796833 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.796869 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.796968 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.797037 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.797068 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.797080 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.797109 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.797154 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.797215 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.797340 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.797423 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.797471 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.797582 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.798092 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.798297 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.798334 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.798362 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.798385 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.798411 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.798434 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.798458 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.798483 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.798508 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.798529 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.798879 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799083 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.798530 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799416 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799451 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799480 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799505 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799532 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799560 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799595 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799595 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799643 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799702 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799735 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799759 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799770 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799817 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799834 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799858 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799881 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799902 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799927 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799949 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799993 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800019 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800040 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800066 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800089 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800114 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800136 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800157 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800175 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800196 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800215 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800236 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800257 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800279 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800301 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800317 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800334 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800332 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800350 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800445 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800458 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800445 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800681 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800693 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800785 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800813 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800820 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800866 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800889 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800933 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800967 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801032 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801066 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801098 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801131 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801140 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801163 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801200 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801256 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801281 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801291 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801335 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801366 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801398 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801431 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801464 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801497 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801534 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801567 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801618 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801656 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801690 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801723 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801764 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801808 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801850 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801896 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801930 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801961 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802020 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802055 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802086 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802117 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802151 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802181 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802215 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802250 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802283 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802314 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802349 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802384 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802416 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802450 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802485 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802519 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802553 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802590 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802622 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802657 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802692 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802725 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802761 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802793 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802825 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802857 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802891 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802923 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802961 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803137 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803194 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803602 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803641 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803679 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803713 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803750 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803785 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803822 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803855 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803889 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803924 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803957 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804020 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804420 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804470 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804506 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804540 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804675 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804714 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804750 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804702 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804785 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804819 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804853 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804887 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804922 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804953 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805010 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805048 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805082 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805117 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805190 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805227 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805263 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805299 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805335 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805368 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805401 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805434 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805471 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805508 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805542 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805573 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805608 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805642 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805675 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805710 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805744 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.807217 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.807408 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.807451 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.807490 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.807574 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.807715 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.807752 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.807788 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.807827 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.807863 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.807896 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808136 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808194 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808238 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808274 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808314 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808352 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808394 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808432 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808467 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808512 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808591 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801491 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808646 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801522 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801539 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801806 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802154 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802411 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802572 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802744 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802936 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808781 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803318 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803665 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803852 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803944 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803963 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804083 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804168 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804343 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804732 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.807192 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.807542 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808103 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808182 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808374 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808712 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808943 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808770 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809095 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809329 4948 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809348 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809363 4948 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809377 4948 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809392 4948 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809404 4948 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809419 4948 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809436 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809451 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809466 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809479 4948 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809491 4948 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809503 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809517 4948 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809531 4948 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809545 4948 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809558 4948 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809572 4948 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809586 4948 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809599 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809613 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809630 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809643 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809657 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809668 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809681 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809694 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809707 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809720 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809732 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809743 4948 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809757 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809769 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809781 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809792 4948 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809804 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809816 4948 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809828 4948 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809840 4948 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809853 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809865 4948 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809877 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809891 4948 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809903 4948 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809916 4948 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809928 4948 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809940 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809952 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809964 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809996 4948 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809437 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809446 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809483 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809509 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809577 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809812 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.810000 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.810097 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.810220 4948 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.810223 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.810288 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:13.310266656 +0000 UTC m=+22.284761606 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.810380 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.810347 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.810455 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.810843 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.810965 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.810958 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.811287 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.811271 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.811598 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.811876 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.811906 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.812081 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.812319 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.812545 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.812737 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.812780 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.812796 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.812152 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.812858 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.813102 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.813181 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.813204 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.813662 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.813582 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.813829 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.814031 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.814317 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.814378 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.814499 4948 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.814706 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.814796 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.814822 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.814989 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.815200 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.815155 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.817361 4948 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.817476 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:13.317443217 +0000 UTC m=+22.291938077 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.817927 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.818253 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.818461 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.821466 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:06:13.321442968 +0000 UTC m=+22.295937778 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.821455 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.830291 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.830670 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.830671 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.831777 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.833155 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.833201 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.833221 4948 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.833668 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.833863 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.834033 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:13.334005295 +0000 UTC m=+22.308500325 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.834266 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.834354 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.834460 4948 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.834588 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:13.334569619 +0000 UTC m=+22.309064439 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.834299 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.834286 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.835167 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.835259 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.836415 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.836928 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.837170 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.837233 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.837644 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.837757 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.837952 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.837963 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.838465 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.839291 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.839771 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.839901 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.842138 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.842868 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.843167 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.843237 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.843285 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.843309 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.843659 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.844026 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.844507 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.844517 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.844605 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.844871 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.845040 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.845159 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.845247 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.845320 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.845517 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.845609 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.845840 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.846103 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.846262 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.846332 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.846343 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.846453 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.846659 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.846879 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.846895 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.847116 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.847236 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.847128 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.847682 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.847813 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.847873 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.847907 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.848033 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.848046 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.848019 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.848051 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.848586 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.848604 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.848806 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.848911 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.848998 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.848949 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.849033 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.849225 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.849462 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.849523 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.849571 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.849538 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.849754 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.850294 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.850652 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.851047 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.850722 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.851177 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.851947 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.853721 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.854180 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.854267 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.855232 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.855286 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.855346 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.855790 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.855942 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.856140 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.855963 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.863424 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.865638 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.882919 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.883863 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.896640 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.898905 4948 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678" exitCode=255 Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.900210 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678"} Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.910479 4948 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.910836 4948 scope.go:117] "RemoveContainer" containerID="65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911200 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911289 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911432 4948 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911452 4948 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911490 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911502 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911513 4948 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911524 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911534 4948 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911565 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911574 4948 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911583 4948 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911591 4948 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911600 4948 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911609 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911617 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911652 4948 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911663 4948 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911677 4948 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911688 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911698 4948 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911727 4948 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911739 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911749 4948 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911758 4948 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911768 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911779 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911805 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911816 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911827 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911841 4948 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911825 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911853 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912111 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912191 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912206 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912218 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912229 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912239 4948 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912249 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912260 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912271 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912281 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912292 4948 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912303 4948 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912349 4948 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912361 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912372 4948 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912385 4948 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912774 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912819 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912829 4948 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912840 4948 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912850 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912859 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912870 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912880 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912889 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912899 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912719 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913190 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913240 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913271 4948 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913298 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913325 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913350 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913375 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913399 4948 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913423 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913447 4948 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913473 4948 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913500 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913529 4948 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913555 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913579 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913605 4948 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913629 4948 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913653 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913676 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913701 4948 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913725 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913752 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913776 4948 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913801 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913883 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913909 4948 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913935 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913955 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914007 4948 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914033 4948 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914059 4948 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914084 4948 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914108 4948 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914133 4948 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914158 4948 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.914171 4948 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-crc\" already exists" pod="openshift-etcd/etcd-crc" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914183 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914209 4948 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914319 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914522 4948 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914539 4948 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914549 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914559 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914569 4948 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914597 4948 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914606 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914615 4948 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914625 4948 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914636 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914648 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914675 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914685 4948 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914695 4948 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914704 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914714 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914726 4948 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914759 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914771 4948 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914782 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914790 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914800 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914900 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914923 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914943 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914963 4948 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915008 4948 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915026 4948 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915044 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915061 4948 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915077 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915094 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915110 4948 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915128 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915145 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915164 4948 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915180 4948 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915199 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915216 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915232 4948 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915249 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915265 4948 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915282 4948 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915299 4948 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915315 4948 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915332 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915349 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915367 4948 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915386 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.932726 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.948898 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.961296 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.964341 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.972066 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.977361 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.977723 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.989537 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:12 crc kubenswrapper[4948]: W0220 08:06:12.994417 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-db483d7683a8ed01afc8e5cfd09fa69c3b88dae193594965ab653e559b96f064 WatchSource:0}: Error finding container db483d7683a8ed01afc8e5cfd09fa69c3b88dae193594965ab653e559b96f064: Status 404 returned error can't find the container with id db483d7683a8ed01afc8e5cfd09fa69c3b88dae193594965ab653e559b96f064 Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.008334 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.018804 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.029415 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.037428 4948 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-20 08:01:12 +0000 UTC, rotation deadline is 2027-01-01 04:33:54.676000435 +0000 UTC Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.037510 4948 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7556h27m41.638495029s for next certificate rotation Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.319644 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.319735 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:13 crc kubenswrapper[4948]: E0220 08:06:13.319866 4948 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 08:06:13 crc kubenswrapper[4948]: E0220 08:06:13.319859 4948 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 08:06:13 crc kubenswrapper[4948]: E0220 08:06:13.319930 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:14.31991459 +0000 UTC m=+23.294409410 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 08:06:13 crc kubenswrapper[4948]: E0220 08:06:13.320020 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:14.319962711 +0000 UTC m=+23.294457561 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.420793 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.420861 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.420905 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:13 crc kubenswrapper[4948]: E0220 08:06:13.421067 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 08:06:13 crc kubenswrapper[4948]: E0220 08:06:13.421087 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 08:06:13 crc kubenswrapper[4948]: E0220 08:06:13.421098 4948 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:13 crc kubenswrapper[4948]: E0220 08:06:13.421172 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:06:14.421103526 +0000 UTC m=+23.395598346 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:06:13 crc kubenswrapper[4948]: E0220 08:06:13.421261 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:14.42124879 +0000 UTC m=+23.395743840 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:13 crc kubenswrapper[4948]: E0220 08:06:13.421352 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 08:06:13 crc kubenswrapper[4948]: E0220 08:06:13.421402 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 08:06:13 crc kubenswrapper[4948]: E0220 08:06:13.421422 4948 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:13 crc kubenswrapper[4948]: E0220 08:06:13.421523 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:14.421493596 +0000 UTC m=+23.395988456 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.664520 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 04:30:43.841625809 +0000 UTC Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.726923 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.727451 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.728396 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.729174 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.729870 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.730531 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.731261 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.731904 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.732771 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.735187 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.736008 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.737301 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.737908 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.739109 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.739761 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.740543 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.742350 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.743273 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.745400 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.746626 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.747637 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.749663 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.750581 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.752786 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.753650 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.755642 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.757571 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.758670 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.760094 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.761223 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.762288 4948 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.762508 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.768688 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.771426 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.772646 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.777259 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.780259 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.781890 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.783530 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.785260 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.786632 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.788404 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.790324 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.793119 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.794591 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.796917 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.799362 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.802151 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.803292 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.804599 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.805779 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.807291 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.808817 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.809596 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.843746 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.848696 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.875197 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.903673 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d"} Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.903802 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"93e398dada94802ff7db7fb1eb7f711e7a157be71ff2d442d8eb788b009b172a"} Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.905141 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.907083 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61"} Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.907208 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.909435 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b"} Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.909463 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237"} Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.909475 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"db483d7683a8ed01afc8e5cfd09fa69c3b88dae193594965ab653e559b96f064"} Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.910455 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"1268deebbef00781e7b186ba2f25f8bfcbfb64830abd8c86bb416b2a6c52d395"} Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.928384 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:13Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.971622 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:13Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.988576 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:13Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.008366 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.026876 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.043572 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.060345 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.081770 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.094253 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-z8rrb"] Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.094561 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-z8rrb" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.102547 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.102735 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.103620 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.120341 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.147674 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.171161 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.185160 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.201474 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.227073 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.228420 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8c9cb414-e3c0-4e40-92b1-42f05565e44c-hosts-file\") pod \"node-resolver-z8rrb\" (UID: \"8c9cb414-e3c0-4e40-92b1-42f05565e44c\") " pod="openshift-dns/node-resolver-z8rrb" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.228476 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tzrg\" (UniqueName: \"kubernetes.io/projected/8c9cb414-e3c0-4e40-92b1-42f05565e44c-kube-api-access-4tzrg\") pod \"node-resolver-z8rrb\" (UID: \"8c9cb414-e3c0-4e40-92b1-42f05565e44c\") " pod="openshift-dns/node-resolver-z8rrb" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.242695 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.253513 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.264399 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.274036 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.286356 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.301524 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.316166 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.328904 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.328946 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8c9cb414-e3c0-4e40-92b1-42f05565e44c-hosts-file\") pod \"node-resolver-z8rrb\" (UID: \"8c9cb414-e3c0-4e40-92b1-42f05565e44c\") " pod="openshift-dns/node-resolver-z8rrb" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.328999 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.329025 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tzrg\" (UniqueName: \"kubernetes.io/projected/8c9cb414-e3c0-4e40-92b1-42f05565e44c-kube-api-access-4tzrg\") pod \"node-resolver-z8rrb\" (UID: \"8c9cb414-e3c0-4e40-92b1-42f05565e44c\") " pod="openshift-dns/node-resolver-z8rrb" Feb 20 08:06:14 crc kubenswrapper[4948]: E0220 08:06:14.329120 4948 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.329174 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8c9cb414-e3c0-4e40-92b1-42f05565e44c-hosts-file\") pod \"node-resolver-z8rrb\" (UID: \"8c9cb414-e3c0-4e40-92b1-42f05565e44c\") " pod="openshift-dns/node-resolver-z8rrb" Feb 20 08:06:14 crc kubenswrapper[4948]: E0220 08:06:14.329209 4948 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 08:06:14 crc kubenswrapper[4948]: E0220 08:06:14.329234 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:16.329209036 +0000 UTC m=+25.303703926 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 08:06:14 crc kubenswrapper[4948]: E0220 08:06:14.329328 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:16.329290518 +0000 UTC m=+25.303785338 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.332054 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.347788 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.359589 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.377498 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.397650 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.422688 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.429903 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.430257 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tzrg\" (UniqueName: \"kubernetes.io/projected/8c9cb414-e3c0-4e40-92b1-42f05565e44c-kube-api-access-4tzrg\") pod \"node-resolver-z8rrb\" (UID: \"8c9cb414-e3c0-4e40-92b1-42f05565e44c\") " pod="openshift-dns/node-resolver-z8rrb" Feb 20 08:06:14 crc kubenswrapper[4948]: E0220 08:06:14.430263 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:06:16.430226737 +0000 UTC m=+25.404721557 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.430494 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.430618 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:14 crc kubenswrapper[4948]: E0220 08:06:14.430722 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 08:06:14 crc kubenswrapper[4948]: E0220 08:06:14.430752 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 08:06:14 crc kubenswrapper[4948]: E0220 08:06:14.430770 4948 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:14 crc kubenswrapper[4948]: E0220 08:06:14.430822 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:16.430804062 +0000 UTC m=+25.405298882 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:14 crc kubenswrapper[4948]: E0220 08:06:14.431055 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 08:06:14 crc kubenswrapper[4948]: E0220 08:06:14.431148 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 08:06:14 crc kubenswrapper[4948]: E0220 08:06:14.431215 4948 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:14 crc kubenswrapper[4948]: E0220 08:06:14.431353 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:16.431326405 +0000 UTC m=+25.405821465 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.665087 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 05:20:23.981600229 +0000 UTC Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.705303 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-z8rrb" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.724197 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.724234 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.724290 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:14 crc kubenswrapper[4948]: E0220 08:06:14.724496 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:14 crc kubenswrapper[4948]: E0220 08:06:14.724333 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:14 crc kubenswrapper[4948]: E0220 08:06:14.724695 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.877942 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-8frp4"] Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.878575 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.879130 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-h4ww2"] Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.879363 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.879424 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-x7s56"] Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.879890 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-x7s56" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.880151 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.881089 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.881597 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.881658 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.882897 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.883336 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.883830 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.884222 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.884259 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.884591 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.884856 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.885413 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.912440 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.920383 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-z8rrb" event={"ID":"8c9cb414-e3c0-4e40-92b1-42f05565e44c","Type":"ContainerStarted","Data":"0482621a1c61d8c484fe2f3b29c02bacf568c4be68044be2ed544076db12509f"} Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.933554 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.957092 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.969047 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.994325 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.007987 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.022663 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.035907 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5757f611-4b0c-49dc-a076-c90ebaad78d8-cni-binary-copy\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.035965 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee-mcd-auth-proxy-config\") pod \"machine-config-daemon-h4ww2\" (UID: \"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\") " pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036009 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b96124e4-0a74-4578-9142-fd728eb9f99e-cni-binary-copy\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036035 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-os-release\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036108 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-var-lib-kubelet\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036159 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b96124e4-0a74-4578-9142-fd728eb9f99e-multus-daemon-config\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036175 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-etc-kubernetes\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036199 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5757f611-4b0c-49dc-a076-c90ebaad78d8-os-release\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036218 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-cnibin\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036234 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5757f611-4b0c-49dc-a076-c90ebaad78d8-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036260 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5757f611-4b0c-49dc-a076-c90ebaad78d8-cnibin\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036300 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5757f611-4b0c-49dc-a076-c90ebaad78d8-tuning-conf-dir\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036364 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-var-lib-cni-multus\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036428 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-hostroot\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036449 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-run-multus-certs\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036473 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9689\" (UniqueName: \"kubernetes.io/projected/5757f611-4b0c-49dc-a076-c90ebaad78d8-kube-api-access-x9689\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036493 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnmlr\" (UniqueName: \"kubernetes.io/projected/84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee-kube-api-access-gnmlr\") pod \"machine-config-daemon-h4ww2\" (UID: \"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\") " pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036508 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-system-cni-dir\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036524 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee-proxy-tls\") pod \"machine-config-daemon-h4ww2\" (UID: \"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\") " pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036540 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-var-lib-cni-bin\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036576 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-multus-socket-dir-parent\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036593 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-multus-cni-dir\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036606 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2fpv\" (UniqueName: \"kubernetes.io/projected/b96124e4-0a74-4578-9142-fd728eb9f99e-kube-api-access-b2fpv\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036624 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-multus-conf-dir\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036661 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee-rootfs\") pod \"machine-config-daemon-h4ww2\" (UID: \"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\") " pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036750 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-run-k8s-cni-cncf-io\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036815 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5757f611-4b0c-49dc-a076-c90ebaad78d8-system-cni-dir\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036842 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-run-netns\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.042760 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.069929 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.099655 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.118546 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.137808 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee-rootfs\") pod \"machine-config-daemon-h4ww2\" (UID: \"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\") " pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.137864 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-run-k8s-cni-cncf-io\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.137896 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5757f611-4b0c-49dc-a076-c90ebaad78d8-system-cni-dir\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.137918 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-run-netns\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.137940 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5757f611-4b0c-49dc-a076-c90ebaad78d8-cni-binary-copy\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.137961 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b96124e4-0a74-4578-9142-fd728eb9f99e-cni-binary-copy\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138001 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee-mcd-auth-proxy-config\") pod \"machine-config-daemon-h4ww2\" (UID: \"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\") " pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.137996 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee-rootfs\") pod \"machine-config-daemon-h4ww2\" (UID: \"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\") " pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138025 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-run-k8s-cni-cncf-io\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138082 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-run-netns\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138052 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-var-lib-kubelet\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138000 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5757f611-4b0c-49dc-a076-c90ebaad78d8-system-cni-dir\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138024 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-var-lib-kubelet\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138150 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b96124e4-0a74-4578-9142-fd728eb9f99e-multus-daemon-config\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138173 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-etc-kubernetes\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138188 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-os-release\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138207 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5757f611-4b0c-49dc-a076-c90ebaad78d8-os-release\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138224 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-cnibin\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138242 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5757f611-4b0c-49dc-a076-c90ebaad78d8-cnibin\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138259 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5757f611-4b0c-49dc-a076-c90ebaad78d8-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138294 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5757f611-4b0c-49dc-a076-c90ebaad78d8-tuning-conf-dir\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138315 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9689\" (UniqueName: \"kubernetes.io/projected/5757f611-4b0c-49dc-a076-c90ebaad78d8-kube-api-access-x9689\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138337 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-var-lib-cni-multus\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138353 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-hostroot\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138366 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-run-multus-certs\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138385 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnmlr\" (UniqueName: \"kubernetes.io/projected/84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee-kube-api-access-gnmlr\") pod \"machine-config-daemon-h4ww2\" (UID: \"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\") " pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138408 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-system-cni-dir\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138427 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-var-lib-cni-bin\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138442 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee-proxy-tls\") pod \"machine-config-daemon-h4ww2\" (UID: \"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\") " pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138466 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-multus-socket-dir-parent\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138483 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-multus-cni-dir\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138496 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2fpv\" (UniqueName: \"kubernetes.io/projected/b96124e4-0a74-4578-9142-fd728eb9f99e-kube-api-access-b2fpv\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138511 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-multus-conf-dir\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138561 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-multus-conf-dir\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138668 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b96124e4-0a74-4578-9142-fd728eb9f99e-cni-binary-copy\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138697 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5757f611-4b0c-49dc-a076-c90ebaad78d8-cni-binary-copy\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138724 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-var-lib-cni-multus\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138742 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-var-lib-cni-bin\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138751 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-etc-kubernetes\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138768 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-hostroot\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138794 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-run-multus-certs\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138837 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee-mcd-auth-proxy-config\") pod \"machine-config-daemon-h4ww2\" (UID: \"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\") " pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138911 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5757f611-4b0c-49dc-a076-c90ebaad78d8-cnibin\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.139065 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-os-release\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.139083 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-system-cni-dir\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.139122 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-multus-cni-dir\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.139127 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5757f611-4b0c-49dc-a076-c90ebaad78d8-os-release\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.139159 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-multus-socket-dir-parent\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.139178 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-cnibin\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.139242 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b96124e4-0a74-4578-9142-fd728eb9f99e-multus-daemon-config\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.139558 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5757f611-4b0c-49dc-a076-c90ebaad78d8-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.139691 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5757f611-4b0c-49dc-a076-c90ebaad78d8-tuning-conf-dir\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.140377 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.143373 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee-proxy-tls\") pod \"machine-config-daemon-h4ww2\" (UID: \"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\") " pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.160336 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2fpv\" (UniqueName: \"kubernetes.io/projected/b96124e4-0a74-4578-9142-fd728eb9f99e-kube-api-access-b2fpv\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.168710 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnmlr\" (UniqueName: \"kubernetes.io/projected/84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee-kube-api-access-gnmlr\") pod \"machine-config-daemon-h4ww2\" (UID: \"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\") " pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.171854 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.180543 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9689\" (UniqueName: \"kubernetes.io/projected/5757f611-4b0c-49dc-a076-c90ebaad78d8-kube-api-access-x9689\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.190265 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.191287 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: W0220 08:06:15.202365 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5757f611_4b0c_49dc_a076_c90ebaad78d8.slice/crio-ea4516e20b5daad40c01344cd4ecddf48b4f75302941031c96ccc00b8502043a WatchSource:0}: Error finding container ea4516e20b5daad40c01344cd4ecddf48b4f75302941031c96ccc00b8502043a: Status 404 returned error can't find the container with id ea4516e20b5daad40c01344cd4ecddf48b4f75302941031c96ccc00b8502043a Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.202349 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.215016 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.221238 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.227179 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.235922 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: W0220 08:06:15.245109 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb96124e4_0a74_4578_9142_fd728eb9f99e.slice/crio-cb3225162882e7895b01638438f3868817f5fc21d94397971005435bdd093c67 WatchSource:0}: Error finding container cb3225162882e7895b01638438f3868817f5fc21d94397971005435bdd093c67: Status 404 returned error can't find the container with id cb3225162882e7895b01638438f3868817f5fc21d94397971005435bdd093c67 Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.252866 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.266433 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7s8gg"] Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.267354 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.271069 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.271660 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.271736 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.271998 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.272512 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.272836 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.278880 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.278998 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.292993 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.306844 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.317327 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.332726 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.340884 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-run-ovn-kubernetes\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.340912 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-slash\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.340927 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-var-lib-openvswitch\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.340951 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-etc-openvswitch\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.340964 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-cni-netd\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.340990 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.341025 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovn-node-metrics-cert\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.341079 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-systemd\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.341121 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-log-socket\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.341149 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtrs5\" (UniqueName: \"kubernetes.io/projected/d6a28bd7-2b83-43f8-b803-bfe41516e071-kube-api-access-gtrs5\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.341186 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-kubelet\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.341204 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-openvswitch\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.341227 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-node-log\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.341246 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-env-overrides\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.341266 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-systemd-units\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.341285 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovnkube-config\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.341310 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-cni-bin\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.341337 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovnkube-script-lib\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.341354 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-run-netns\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.341388 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-ovn\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.349893 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.371516 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.387606 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.400215 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.414053 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.425998 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442536 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-etc-openvswitch\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442580 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-cni-netd\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442598 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442671 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovn-node-metrics-cert\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442693 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-systemd\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442709 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtrs5\" (UniqueName: \"kubernetes.io/projected/d6a28bd7-2b83-43f8-b803-bfe41516e071-kube-api-access-gtrs5\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442734 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-log-socket\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442771 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-kubelet\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442790 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-openvswitch\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442811 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-node-log\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442827 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-env-overrides\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442847 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-systemd-units\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442863 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovnkube-config\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442881 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-cni-bin\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442897 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovnkube-script-lib\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442922 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-run-netns\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442942 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-ovn\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442960 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-run-ovn-kubernetes\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443001 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-slash\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443019 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-var-lib-openvswitch\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443091 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-var-lib-openvswitch\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443132 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-etc-openvswitch\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443156 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-cni-netd\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443178 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443203 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-systemd-units\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443225 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-ovn\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443248 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-run-ovn-kubernetes\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443301 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-run-netns\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443359 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-systemd\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443377 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-openvswitch\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443387 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-kubelet\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443391 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-cni-bin\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443406 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-node-log\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443393 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-slash\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443540 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-log-socket\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443907 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-env-overrides\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.444203 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovnkube-script-lib\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.444355 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovnkube-config\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.448921 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.449143 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovn-node-metrics-cert\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.464556 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtrs5\" (UniqueName: \"kubernetes.io/projected/d6a28bd7-2b83-43f8-b803-bfe41516e071-kube-api-access-gtrs5\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.465828 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.482547 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.498928 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.512247 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.525712 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.538301 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.550503 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.562931 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.665376 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 10:31:20.104008261 +0000 UTC Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.673733 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: W0220 08:06:15.685485 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6a28bd7_2b83_43f8_b803_bfe41516e071.slice/crio-0269d0e4618f9bfc08bbd4b396cc89a6bd78be17afe5c9425cf43f02da885bc5 WatchSource:0}: Error finding container 0269d0e4618f9bfc08bbd4b396cc89a6bd78be17afe5c9425cf43f02da885bc5: Status 404 returned error can't find the container with id 0269d0e4618f9bfc08bbd4b396cc89a6bd78be17afe5c9425cf43f02da885bc5 Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.926443 4948 generic.go:334] "Generic (PLEG): container finished" podID="5757f611-4b0c-49dc-a076-c90ebaad78d8" containerID="dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069" exitCode=0 Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.926522 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" event={"ID":"5757f611-4b0c-49dc-a076-c90ebaad78d8","Type":"ContainerDied","Data":"dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069"} Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.926559 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" event={"ID":"5757f611-4b0c-49dc-a076-c90ebaad78d8","Type":"ContainerStarted","Data":"ea4516e20b5daad40c01344cd4ecddf48b4f75302941031c96ccc00b8502043a"} Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.928320 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x7s56" event={"ID":"b96124e4-0a74-4578-9142-fd728eb9f99e","Type":"ContainerStarted","Data":"34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e"} Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.928350 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x7s56" event={"ID":"b96124e4-0a74-4578-9142-fd728eb9f99e","Type":"ContainerStarted","Data":"cb3225162882e7895b01638438f3868817f5fc21d94397971005435bdd093c67"} Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.934749 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d"} Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.937770 4948 generic.go:334] "Generic (PLEG): container finished" podID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerID="540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad" exitCode=0 Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.937860 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerDied","Data":"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad"} Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.938036 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerStarted","Data":"0269d0e4618f9bfc08bbd4b396cc89a6bd78be17afe5c9425cf43f02da885bc5"} Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.940180 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e"} Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.940206 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59"} Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.940217 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"d824cc19743acd2484c541a11dd55f5e646b98af34783b9779359153056f8cb3"} Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.942217 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-z8rrb" event={"ID":"8c9cb414-e3c0-4e40-92b1-42f05565e44c","Type":"ContainerStarted","Data":"180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b"} Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.960422 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.977260 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.007516 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.024608 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.037540 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.053395 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.068282 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.087690 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.102849 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.120560 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.134222 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.145138 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.180893 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.197952 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.224886 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.236691 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.251304 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.265016 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.288395 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.301391 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.322414 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.337587 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.352037 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.354528 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.354577 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:16 crc kubenswrapper[4948]: E0220 08:06:16.354702 4948 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 08:06:16 crc kubenswrapper[4948]: E0220 08:06:16.354708 4948 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 08:06:16 crc kubenswrapper[4948]: E0220 08:06:16.354771 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:20.354752592 +0000 UTC m=+29.329247412 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 08:06:16 crc kubenswrapper[4948]: E0220 08:06:16.354809 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:20.354783963 +0000 UTC m=+29.329278793 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.366090 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.386930 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.401678 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.417164 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.433334 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.455736 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:06:16 crc kubenswrapper[4948]: E0220 08:06:16.455887 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:06:20.455860779 +0000 UTC m=+29.430355599 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.455946 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:16 crc kubenswrapper[4948]: E0220 08:06:16.456101 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 08:06:16 crc kubenswrapper[4948]: E0220 08:06:16.456131 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 08:06:16 crc kubenswrapper[4948]: E0220 08:06:16.456144 4948 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.456137 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:16 crc kubenswrapper[4948]: E0220 08:06:16.456181 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:20.456171786 +0000 UTC m=+29.430666736 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:16 crc kubenswrapper[4948]: E0220 08:06:16.456309 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 08:06:16 crc kubenswrapper[4948]: E0220 08:06:16.456343 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 08:06:16 crc kubenswrapper[4948]: E0220 08:06:16.456358 4948 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:16 crc kubenswrapper[4948]: E0220 08:06:16.456418 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:20.456399372 +0000 UTC m=+29.430894242 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.666043 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 12:39:18.947377048 +0000 UTC Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.721612 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.721671 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.721735 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:16 crc kubenswrapper[4948]: E0220 08:06:16.721774 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:16 crc kubenswrapper[4948]: E0220 08:06:16.721957 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:16 crc kubenswrapper[4948]: E0220 08:06:16.722179 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.950727 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerStarted","Data":"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2"} Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.951055 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerStarted","Data":"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37"} Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.951072 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerStarted","Data":"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c"} Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.951087 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerStarted","Data":"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7"} Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.951100 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerStarted","Data":"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d"} Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.953015 4948 generic.go:334] "Generic (PLEG): container finished" podID="5757f611-4b0c-49dc-a076-c90ebaad78d8" containerID="9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32" exitCode=0 Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.953102 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" event={"ID":"5757f611-4b0c-49dc-a076-c90ebaad78d8","Type":"ContainerDied","Data":"9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32"} Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.976173 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.997375 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.014738 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.043436 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.069395 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.085716 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.107666 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.132203 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.157909 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.158390 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-nsk7l"] Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.158862 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-nsk7l" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.160715 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.161037 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.161175 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.161475 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.171761 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.189637 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.205573 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.219902 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.236185 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.261330 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.267617 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dc562\" (UniqueName: \"kubernetes.io/projected/ba900a81-fc7b-40b3-942c-097c47812bf4-kube-api-access-dc562\") pod \"node-ca-nsk7l\" (UID: \"ba900a81-fc7b-40b3-942c-097c47812bf4\") " pod="openshift-image-registry/node-ca-nsk7l" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.267756 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/ba900a81-fc7b-40b3-942c-097c47812bf4-serviceca\") pod \"node-ca-nsk7l\" (UID: \"ba900a81-fc7b-40b3-942c-097c47812bf4\") " pod="openshift-image-registry/node-ca-nsk7l" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.267815 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ba900a81-fc7b-40b3-942c-097c47812bf4-host\") pod \"node-ca-nsk7l\" (UID: \"ba900a81-fc7b-40b3-942c-097c47812bf4\") " pod="openshift-image-registry/node-ca-nsk7l" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.276299 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.294285 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.307596 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.334081 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.349270 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.362547 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.369177 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ba900a81-fc7b-40b3-942c-097c47812bf4-host\") pod \"node-ca-nsk7l\" (UID: \"ba900a81-fc7b-40b3-942c-097c47812bf4\") " pod="openshift-image-registry/node-ca-nsk7l" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.369347 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ba900a81-fc7b-40b3-942c-097c47812bf4-host\") pod \"node-ca-nsk7l\" (UID: \"ba900a81-fc7b-40b3-942c-097c47812bf4\") " pod="openshift-image-registry/node-ca-nsk7l" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.369270 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dc562\" (UniqueName: \"kubernetes.io/projected/ba900a81-fc7b-40b3-942c-097c47812bf4-kube-api-access-dc562\") pod \"node-ca-nsk7l\" (UID: \"ba900a81-fc7b-40b3-942c-097c47812bf4\") " pod="openshift-image-registry/node-ca-nsk7l" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.369484 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/ba900a81-fc7b-40b3-942c-097c47812bf4-serviceca\") pod \"node-ca-nsk7l\" (UID: \"ba900a81-fc7b-40b3-942c-097c47812bf4\") " pod="openshift-image-registry/node-ca-nsk7l" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.370601 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/ba900a81-fc7b-40b3-942c-097c47812bf4-serviceca\") pod \"node-ca-nsk7l\" (UID: \"ba900a81-fc7b-40b3-942c-097c47812bf4\") " pod="openshift-image-registry/node-ca-nsk7l" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.385883 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.399745 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dc562\" (UniqueName: \"kubernetes.io/projected/ba900a81-fc7b-40b3-942c-097c47812bf4-kube-api-access-dc562\") pod \"node-ca-nsk7l\" (UID: \"ba900a81-fc7b-40b3-942c-097c47812bf4\") " pod="openshift-image-registry/node-ca-nsk7l" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.405132 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.416760 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.432459 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.445766 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.460792 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.471647 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-nsk7l" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.471700 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.484777 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.666747 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 11:35:45.396778704 +0000 UTC Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.960717 4948 generic.go:334] "Generic (PLEG): container finished" podID="5757f611-4b0c-49dc-a076-c90ebaad78d8" containerID="c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6" exitCode=0 Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.960799 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" event={"ID":"5757f611-4b0c-49dc-a076-c90ebaad78d8","Type":"ContainerDied","Data":"c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6"} Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.966229 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerStarted","Data":"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a"} Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.967744 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-nsk7l" event={"ID":"ba900a81-fc7b-40b3-942c-097c47812bf4","Type":"ContainerStarted","Data":"eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7"} Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.967796 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-nsk7l" event={"ID":"ba900a81-fc7b-40b3-942c-097c47812bf4","Type":"ContainerStarted","Data":"22fe034e4c89f89b70792ebcca37ec652f2177605c3edc3505d40246c825794e"} Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.987603 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.007311 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.037632 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.060671 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.076490 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.099027 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.121767 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.141719 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.162436 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.186086 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.201655 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.219537 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.248627 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.266697 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.282950 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.304657 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.318936 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.337554 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.356769 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.388820 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.391611 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.391667 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.391684 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.391827 4948 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.403205 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.406877 4948 kubelet_node_status.go:115] "Node was previously registered" node="crc" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.407305 4948 kubelet_node_status.go:79] "Successfully registered node" node="crc" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.408736 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.408779 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.408798 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.408820 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.408837 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:18Z","lastTransitionTime":"2026-02-20T08:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.427608 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.443200 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: E0220 08:06:18.448871 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.456057 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.456101 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.456113 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.456131 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.456143 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:18Z","lastTransitionTime":"2026-02-20T08:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.463657 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: E0220 08:06:18.469232 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.472509 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.472584 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.472601 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.472626 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.472642 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:18Z","lastTransitionTime":"2026-02-20T08:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.483052 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: E0220 08:06:18.486737 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.491107 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.491162 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.491174 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.491196 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.491224 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:18Z","lastTransitionTime":"2026-02-20T08:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.501369 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: E0220 08:06:18.504525 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.509597 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.509638 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.509686 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.509710 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.509725 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:18Z","lastTransitionTime":"2026-02-20T08:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.510940 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: E0220 08:06:18.520571 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: E0220 08:06:18.520692 4948 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.522304 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.522331 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.522342 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.522360 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.522373 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:18Z","lastTransitionTime":"2026-02-20T08:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.551704 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.590946 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.626065 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.626113 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.626125 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.626142 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.626154 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:18Z","lastTransitionTime":"2026-02-20T08:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.627499 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.664957 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.667306 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 00:39:35.321571782 +0000 UTC Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.722280 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.722408 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:18 crc kubenswrapper[4948]: E0220 08:06:18.722527 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.722575 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:18 crc kubenswrapper[4948]: E0220 08:06:18.722765 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:18 crc kubenswrapper[4948]: E0220 08:06:18.722870 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.729657 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.729715 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.729733 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.729786 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.729806 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:18Z","lastTransitionTime":"2026-02-20T08:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.833425 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.833482 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.833499 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.833521 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.833540 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:18Z","lastTransitionTime":"2026-02-20T08:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.943034 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.943096 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.943114 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.943139 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.943158 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:18Z","lastTransitionTime":"2026-02-20T08:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.977052 4948 generic.go:334] "Generic (PLEG): container finished" podID="5757f611-4b0c-49dc-a076-c90ebaad78d8" containerID="b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155" exitCode=0 Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.977111 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" event={"ID":"5757f611-4b0c-49dc-a076-c90ebaad78d8","Type":"ContainerDied","Data":"b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155"} Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.000573 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.026153 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:19Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.047298 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.047348 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.047366 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.047391 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.047410 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:19Z","lastTransitionTime":"2026-02-20T08:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.051518 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:19Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.077045 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:19Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.095014 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:19Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.112184 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:19Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.129938 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:19Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.157668 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.157735 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.157748 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.157774 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.157791 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:19Z","lastTransitionTime":"2026-02-20T08:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.162243 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:19Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.185260 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:19Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.198319 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:19Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.227427 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:19Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.244967 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:19Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.260761 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.260802 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.260812 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.260830 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.260845 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:19Z","lastTransitionTime":"2026-02-20T08:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.271459 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:19Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.289540 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:19Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.305821 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:19Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.363576 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.363620 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.363632 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.363652 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.363666 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:19Z","lastTransitionTime":"2026-02-20T08:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.467182 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.467257 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.467288 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.467321 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.467340 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:19Z","lastTransitionTime":"2026-02-20T08:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.570952 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.571032 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.571045 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.571066 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.571078 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:19Z","lastTransitionTime":"2026-02-20T08:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.667834 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 11:54:04.715149774 +0000 UTC Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.675325 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.675371 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.675384 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.675404 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.675419 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:19Z","lastTransitionTime":"2026-02-20T08:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.779109 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.779190 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.779209 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.779237 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.779256 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:19Z","lastTransitionTime":"2026-02-20T08:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.882199 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.882271 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.882291 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.882319 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.882341 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:19Z","lastTransitionTime":"2026-02-20T08:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.990603 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.990660 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.990681 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.990708 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.990734 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:19Z","lastTransitionTime":"2026-02-20T08:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.995261 4948 generic.go:334] "Generic (PLEG): container finished" podID="5757f611-4b0c-49dc-a076-c90ebaad78d8" containerID="1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc" exitCode=0 Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.995345 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" event={"ID":"5757f611-4b0c-49dc-a076-c90ebaad78d8","Type":"ContainerDied","Data":"1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc"} Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.006362 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerStarted","Data":"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022"} Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.035293 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.060217 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.080300 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.093924 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.094012 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.094031 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.094058 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.094079 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:20Z","lastTransitionTime":"2026-02-20T08:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.101554 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.131324 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.156412 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.175189 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.199053 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.199154 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.199220 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.199251 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.199317 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:20Z","lastTransitionTime":"2026-02-20T08:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.205483 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.227358 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.253321 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.274269 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.298487 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.302909 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.302997 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.303010 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.303039 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.303055 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:20Z","lastTransitionTime":"2026-02-20T08:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.319478 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.337960 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.359279 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.403684 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.403763 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:20 crc kubenswrapper[4948]: E0220 08:06:20.403947 4948 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 08:06:20 crc kubenswrapper[4948]: E0220 08:06:20.404078 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:28.404054763 +0000 UTC m=+37.378549603 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 08:06:20 crc kubenswrapper[4948]: E0220 08:06:20.404647 4948 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 08:06:20 crc kubenswrapper[4948]: E0220 08:06:20.404727 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:28.404708739 +0000 UTC m=+37.379203569 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.407112 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.407153 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.407166 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.407185 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.407199 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:20Z","lastTransitionTime":"2026-02-20T08:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.504894 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:06:20 crc kubenswrapper[4948]: E0220 08:06:20.505177 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:06:28.505139375 +0000 UTC m=+37.479634235 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.505334 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.505453 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:20 crc kubenswrapper[4948]: E0220 08:06:20.505617 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 08:06:20 crc kubenswrapper[4948]: E0220 08:06:20.505645 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 08:06:20 crc kubenswrapper[4948]: E0220 08:06:20.505660 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 08:06:20 crc kubenswrapper[4948]: E0220 08:06:20.505674 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 08:06:20 crc kubenswrapper[4948]: E0220 08:06:20.505688 4948 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:20 crc kubenswrapper[4948]: E0220 08:06:20.505695 4948 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:20 crc kubenswrapper[4948]: E0220 08:06:20.505760 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:28.505739869 +0000 UTC m=+37.480234719 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:20 crc kubenswrapper[4948]: E0220 08:06:20.505786 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:28.50577459 +0000 UTC m=+37.480269450 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.510613 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.510670 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.510686 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.510712 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.510732 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:20Z","lastTransitionTime":"2026-02-20T08:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.614001 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.614072 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.614088 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.614111 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.614128 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:20Z","lastTransitionTime":"2026-02-20T08:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.668359 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 03:45:56.265401649 +0000 UTC Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.716622 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.716676 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.716691 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.716708 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.716722 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:20Z","lastTransitionTime":"2026-02-20T08:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.722074 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.722074 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:20 crc kubenswrapper[4948]: E0220 08:06:20.722195 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:20 crc kubenswrapper[4948]: E0220 08:06:20.722271 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.722093 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:20 crc kubenswrapper[4948]: E0220 08:06:20.722341 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.820163 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.820208 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.820220 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.820237 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.820249 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:20Z","lastTransitionTime":"2026-02-20T08:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.922820 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.922916 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.922939 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.922968 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.923026 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:20Z","lastTransitionTime":"2026-02-20T08:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.014658 4948 generic.go:334] "Generic (PLEG): container finished" podID="5757f611-4b0c-49dc-a076-c90ebaad78d8" containerID="86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4" exitCode=0 Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.014730 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" event={"ID":"5757f611-4b0c-49dc-a076-c90ebaad78d8","Type":"ContainerDied","Data":"86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4"} Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.025452 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.025516 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.025537 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.025560 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.025578 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:21Z","lastTransitionTime":"2026-02-20T08:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.038768 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.061453 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.075459 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.094900 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.117133 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.130614 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.130687 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.130712 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.130743 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.130766 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:21Z","lastTransitionTime":"2026-02-20T08:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.134007 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.148123 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.167773 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.182035 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.198393 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.210379 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.234185 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.234237 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.234249 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.234268 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.234283 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:21Z","lastTransitionTime":"2026-02-20T08:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.235848 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.256884 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.274656 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.291268 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.337645 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.337700 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.337718 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.337745 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.337768 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:21Z","lastTransitionTime":"2026-02-20T08:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.440349 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.440439 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.440507 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.440541 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.440564 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:21Z","lastTransitionTime":"2026-02-20T08:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.487160 4948 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.543469 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.543658 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.543759 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.544029 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.544113 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:21Z","lastTransitionTime":"2026-02-20T08:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.647427 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.647496 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.647525 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.647559 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.647583 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:21Z","lastTransitionTime":"2026-02-20T08:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.669123 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 23:46:06.816799505 +0000 UTC Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.750756 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.750818 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.750837 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.750862 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.750881 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:21Z","lastTransitionTime":"2026-02-20T08:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.752258 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.775145 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.795726 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.828038 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.854747 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.854818 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.854842 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.854874 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.854898 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:21Z","lastTransitionTime":"2026-02-20T08:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.858440 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.880641 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.901628 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.928413 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.952616 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.958060 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.958105 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.958124 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.958150 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.958171 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:21Z","lastTransitionTime":"2026-02-20T08:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.967659 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.997967 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.017086 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.023504 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" event={"ID":"5757f611-4b0c-49dc-a076-c90ebaad78d8","Type":"ContainerStarted","Data":"78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c"} Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.031323 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerStarted","Data":"3456b89afa02fb3c4b83afea773783e0ae1c65903d24789d1a742688b7745b1a"} Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.032199 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.036743 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.061367 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.061431 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.061447 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.061474 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.061494 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:22Z","lastTransitionTime":"2026-02-20T08:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.063557 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.081246 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.085416 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.098141 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.116582 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.134227 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.165307 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.165356 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.165373 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.165396 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.165416 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:22Z","lastTransitionTime":"2026-02-20T08:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.169452 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3456b89afa02fb3c4b83afea773783e0ae1c65903d24789d1a742688b7745b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.188392 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.207287 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.228263 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.249531 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.269393 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.269460 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.269485 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.269515 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.269539 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:22Z","lastTransitionTime":"2026-02-20T08:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.270569 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.289762 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.311813 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.329496 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.347210 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.373373 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.373416 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.373445 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.373464 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.373475 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:22Z","lastTransitionTime":"2026-02-20T08:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.384885 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.405298 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.424657 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.441220 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.455109 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.470131 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.476160 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.476229 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.476239 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.476257 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.476269 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:22Z","lastTransitionTime":"2026-02-20T08:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.504192 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.522546 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.537310 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.557341 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.573532 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.579045 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.579095 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.579114 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.579140 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.579159 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:22Z","lastTransitionTime":"2026-02-20T08:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.592562 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.607717 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.637584 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3456b89afa02fb3c4b83afea773783e0ae1c65903d24789d1a742688b7745b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.659343 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.669708 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 07:19:09.782432833 +0000 UTC Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.676324 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.682078 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.682153 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.682183 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.682210 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.682227 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:22Z","lastTransitionTime":"2026-02-20T08:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.693842 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.722099 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.722099 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:22 crc kubenswrapper[4948]: E0220 08:06:22.722261 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:22 crc kubenswrapper[4948]: E0220 08:06:22.722422 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.722723 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:22 crc kubenswrapper[4948]: E0220 08:06:22.723041 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.784437 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.784716 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.784802 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.784884 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.784990 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:22Z","lastTransitionTime":"2026-02-20T08:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.888370 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.888731 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.888941 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.889187 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.889323 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:22Z","lastTransitionTime":"2026-02-20T08:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.993040 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.993105 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.993128 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.993160 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.993182 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:22Z","lastTransitionTime":"2026-02-20T08:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.034897 4948 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.035627 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.070388 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.094643 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:23Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.097075 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.097142 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.097162 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.097193 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.097210 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:23Z","lastTransitionTime":"2026-02-20T08:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.111167 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:23Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.136507 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:23Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.155937 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:23Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.188164 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3456b89afa02fb3c4b83afea773783e0ae1c65903d24789d1a742688b7745b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:23Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.200424 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.200861 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.200881 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.200906 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.200927 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:23Z","lastTransitionTime":"2026-02-20T08:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.211826 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:23Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.232669 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:23Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.253046 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:23Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.273850 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:23Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.300040 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:23Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.304366 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.304436 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.304461 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.304491 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.304517 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:23Z","lastTransitionTime":"2026-02-20T08:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.327025 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:23Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.358376 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:23Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.393903 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:23Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.407619 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.407697 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.407722 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.407758 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.407781 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:23Z","lastTransitionTime":"2026-02-20T08:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.417730 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:23Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.435316 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:23Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.510598 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.510663 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.510685 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.510710 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.510728 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:23Z","lastTransitionTime":"2026-02-20T08:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.613550 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.613614 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.613632 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.613663 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.613682 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:23Z","lastTransitionTime":"2026-02-20T08:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.670876 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 10:33:34.186661742 +0000 UTC Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.716400 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.716464 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.716482 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.716506 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.716527 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:23Z","lastTransitionTime":"2026-02-20T08:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.820526 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.820579 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.820590 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.820608 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.820619 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:23Z","lastTransitionTime":"2026-02-20T08:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.923684 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.923735 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.923746 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.923765 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.923777 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:23Z","lastTransitionTime":"2026-02-20T08:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.026501 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.026559 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.026579 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.026605 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.026623 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:24Z","lastTransitionTime":"2026-02-20T08:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.037737 4948 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.129487 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.129555 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.129573 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.129599 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.129617 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:24Z","lastTransitionTime":"2026-02-20T08:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.232433 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.232507 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.232522 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.232541 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.232580 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:24Z","lastTransitionTime":"2026-02-20T08:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.336036 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.336071 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.336080 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.336095 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.336105 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:24Z","lastTransitionTime":"2026-02-20T08:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.443880 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.444218 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.444328 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.445350 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.445441 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:24Z","lastTransitionTime":"2026-02-20T08:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.548706 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.548767 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.548785 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.548814 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.548838 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:24Z","lastTransitionTime":"2026-02-20T08:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.652039 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.652129 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.652149 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.652175 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.652194 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:24Z","lastTransitionTime":"2026-02-20T08:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.672853 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 05:25:01.737493783 +0000 UTC Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.722366 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.722440 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.722436 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:24 crc kubenswrapper[4948]: E0220 08:06:24.722569 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:24 crc kubenswrapper[4948]: E0220 08:06:24.722655 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:24 crc kubenswrapper[4948]: E0220 08:06:24.722761 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.755156 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.755251 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.755296 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.755325 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.755340 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:24Z","lastTransitionTime":"2026-02-20T08:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.858319 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.858402 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.858425 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.858454 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.858474 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:24Z","lastTransitionTime":"2026-02-20T08:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.961144 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.961216 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.961235 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.961261 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.961280 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:24Z","lastTransitionTime":"2026-02-20T08:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.045181 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovnkube-controller/0.log" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.048935 4948 generic.go:334] "Generic (PLEG): container finished" podID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerID="3456b89afa02fb3c4b83afea773783e0ae1c65903d24789d1a742688b7745b1a" exitCode=1 Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.049045 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerDied","Data":"3456b89afa02fb3c4b83afea773783e0ae1c65903d24789d1a742688b7745b1a"} Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.050191 4948 scope.go:117] "RemoveContainer" containerID="3456b89afa02fb3c4b83afea773783e0ae1c65903d24789d1a742688b7745b1a" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.064671 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.064758 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.064787 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.064819 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.064845 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:25Z","lastTransitionTime":"2026-02-20T08:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.089598 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:25Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.114878 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:25Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.131083 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:25Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.155845 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:25Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.167498 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.167555 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.167575 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.167600 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.167619 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:25Z","lastTransitionTime":"2026-02-20T08:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.174367 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:25Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.209836 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3456b89afa02fb3c4b83afea773783e0ae1c65903d24789d1a742688b7745b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3456b89afa02fb3c4b83afea773783e0ae1c65903d24789d1a742688b7745b1a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:24Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI0220 08:06:24.465835 6302 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:06:24.466938 6302 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0220 08:06:24.467064 6302 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 08:06:24.467092 6302 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0220 08:06:24.467102 6302 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0220 08:06:24.467169 6302 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 08:06:24.467179 6302 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 08:06:24.467254 6302 handler.go:208] Removed *v1.Node event handler 2\\\\nI0220 08:06:24.467267 6302 factory.go:656] Stopping watch factory\\\\nI0220 08:06:24.467295 6302 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 08:06:24.467300 6302 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0220 08:06:24.467312 6302 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 08:06:24.467326 6302 handler.go:208] Removed *v1.Node event handler 7\\\\nI0220 08:06:24.467329 6302 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:25Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.237349 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:25Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.257878 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:25Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.270548 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.270602 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.270622 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.270666 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.270686 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:25Z","lastTransitionTime":"2026-02-20T08:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.279617 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:25Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.303008 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:25Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.321387 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:25Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.344358 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:25Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.362098 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:25Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.373903 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.374016 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.374037 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.374063 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.374548 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:25Z","lastTransitionTime":"2026-02-20T08:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.381408 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:25Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.402025 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:25Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.478284 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.478330 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.478342 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.478362 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.478376 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:25Z","lastTransitionTime":"2026-02-20T08:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.581268 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.581317 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.581326 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.581344 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.581355 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:25Z","lastTransitionTime":"2026-02-20T08:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.673015 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 06:41:28.43946064 +0000 UTC Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.684471 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.684536 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.684561 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.684596 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.684622 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:25Z","lastTransitionTime":"2026-02-20T08:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.788003 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.788063 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.788083 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.788106 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.788130 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:25Z","lastTransitionTime":"2026-02-20T08:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.891158 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.891227 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.891242 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.891268 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.891282 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:25Z","lastTransitionTime":"2026-02-20T08:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.995201 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.995271 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.995289 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.995313 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.995332 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:25Z","lastTransitionTime":"2026-02-20T08:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.058001 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovnkube-controller/0.log" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.062769 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerStarted","Data":"8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e"} Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.062894 4948 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.098643 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.098855 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.099033 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.099195 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.099317 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:26Z","lastTransitionTime":"2026-02-20T08:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.104181 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:26Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.126087 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:26Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.143220 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:26Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.164061 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:26Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.179458 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:26Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.201844 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:26Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.202558 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.202624 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.202643 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.202668 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.202688 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:26Z","lastTransitionTime":"2026-02-20T08:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.220381 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:26Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.252145 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3456b89afa02fb3c4b83afea773783e0ae1c65903d24789d1a742688b7745b1a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:24Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI0220 08:06:24.465835 6302 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:06:24.466938 6302 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0220 08:06:24.467064 6302 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 08:06:24.467092 6302 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0220 08:06:24.467102 6302 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0220 08:06:24.467169 6302 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 08:06:24.467179 6302 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 08:06:24.467254 6302 handler.go:208] Removed *v1.Node event handler 2\\\\nI0220 08:06:24.467267 6302 factory.go:656] Stopping watch factory\\\\nI0220 08:06:24.467295 6302 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 08:06:24.467300 6302 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0220 08:06:24.467312 6302 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 08:06:24.467326 6302 handler.go:208] Removed *v1.Node event handler 7\\\\nI0220 08:06:24.467329 6302 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:26Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.276108 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:26Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.299492 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:26Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.305423 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.305476 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.305492 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.305518 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.305536 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:26Z","lastTransitionTime":"2026-02-20T08:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.321097 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:26Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.347915 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:26Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.370409 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:26Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.390278 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:26Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.408390 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.408458 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.408482 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.408513 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.408541 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:26Z","lastTransitionTime":"2026-02-20T08:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.413096 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:26Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.511397 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.511439 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.511450 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.511466 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.511478 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:26Z","lastTransitionTime":"2026-02-20T08:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.615021 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.615098 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.615123 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.615154 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.615180 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:26Z","lastTransitionTime":"2026-02-20T08:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.674292 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 12:05:26.593601973 +0000 UTC Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.718235 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.718328 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.718348 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.718374 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.718393 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:26Z","lastTransitionTime":"2026-02-20T08:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.721611 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.721640 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.721610 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:26 crc kubenswrapper[4948]: E0220 08:06:26.721772 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:26 crc kubenswrapper[4948]: E0220 08:06:26.721921 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:26 crc kubenswrapper[4948]: E0220 08:06:26.722134 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.821799 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.821928 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.821958 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.822036 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.822063 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:26Z","lastTransitionTime":"2026-02-20T08:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.924954 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.925387 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.925521 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.925622 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.925730 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:26Z","lastTransitionTime":"2026-02-20T08:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.028839 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.028896 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.028913 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.028938 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.028960 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:27Z","lastTransitionTime":"2026-02-20T08:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.067861 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovnkube-controller/1.log" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.068444 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovnkube-controller/0.log" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.071775 4948 generic.go:334] "Generic (PLEG): container finished" podID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerID="8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e" exitCode=1 Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.071831 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerDied","Data":"8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e"} Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.071898 4948 scope.go:117] "RemoveContainer" containerID="3456b89afa02fb3c4b83afea773783e0ae1c65903d24789d1a742688b7745b1a" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.072863 4948 scope.go:117] "RemoveContainer" containerID="8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e" Feb 20 08:06:27 crc kubenswrapper[4948]: E0220 08:06:27.073135 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.091033 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.107040 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.130622 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.132120 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.132170 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.132188 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.132213 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.132237 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:27Z","lastTransitionTime":"2026-02-20T08:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.149854 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.180368 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3456b89afa02fb3c4b83afea773783e0ae1c65903d24789d1a742688b7745b1a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:24Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI0220 08:06:24.465835 6302 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:06:24.466938 6302 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0220 08:06:24.467064 6302 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 08:06:24.467092 6302 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0220 08:06:24.467102 6302 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0220 08:06:24.467169 6302 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 08:06:24.467179 6302 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 08:06:24.467254 6302 handler.go:208] Removed *v1.Node event handler 2\\\\nI0220 08:06:24.467267 6302 factory.go:656] Stopping watch factory\\\\nI0220 08:06:24.467295 6302 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 08:06:24.467300 6302 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0220 08:06:24.467312 6302 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 08:06:24.467326 6302 handler.go:208] Removed *v1.Node event handler 7\\\\nI0220 08:06:24.467329 6302 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:26Z\\\",\\\"message\\\":\\\"rmers/externalversions/factory.go:141\\\\nI0220 08:06:26.064417 6421 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:06:26.065536 6421 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0220 08:06:26.065606 6421 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 08:06:26.065641 6421 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 08:06:26.065650 6421 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 08:06:26.065682 6421 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0220 08:06:26.065678 6421 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0220 08:06:26.065713 6421 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 08:06:26.065690 6421 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0220 08:06:26.065709 6421 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 08:06:26.065775 6421 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 08:06:26.065788 6421 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0220 08:06:26.065794 6421 factory.go:656] Stopping watch factory\\\\nI0220 08:06:26.065810 6421 handler.go:208] Removed *v1.Pod ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.202770 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.224061 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.235507 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.235578 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.235600 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.235628 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.235647 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:27Z","lastTransitionTime":"2026-02-20T08:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.245288 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.266103 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.288393 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.301331 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.308604 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.327325 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.341026 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.341107 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.341127 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.341568 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.341622 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:27Z","lastTransitionTime":"2026-02-20T08:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.358906 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.377757 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.392455 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.414934 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.430555 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.445322 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.445407 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.445432 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.445466 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.445491 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:27Z","lastTransitionTime":"2026-02-20T08:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.469058 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.491672 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.507686 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.531366 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.548955 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.549058 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.549078 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.549105 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.549129 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:27Z","lastTransitionTime":"2026-02-20T08:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.551906 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.587080 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3456b89afa02fb3c4b83afea773783e0ae1c65903d24789d1a742688b7745b1a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:24Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI0220 08:06:24.465835 6302 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:06:24.466938 6302 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0220 08:06:24.467064 6302 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 08:06:24.467092 6302 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0220 08:06:24.467102 6302 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0220 08:06:24.467169 6302 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 08:06:24.467179 6302 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 08:06:24.467254 6302 handler.go:208] Removed *v1.Node event handler 2\\\\nI0220 08:06:24.467267 6302 factory.go:656] Stopping watch factory\\\\nI0220 08:06:24.467295 6302 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 08:06:24.467300 6302 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0220 08:06:24.467312 6302 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 08:06:24.467326 6302 handler.go:208] Removed *v1.Node event handler 7\\\\nI0220 08:06:24.467329 6302 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:26Z\\\",\\\"message\\\":\\\"rmers/externalversions/factory.go:141\\\\nI0220 08:06:26.064417 6421 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:06:26.065536 6421 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0220 08:06:26.065606 6421 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 08:06:26.065641 6421 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 08:06:26.065650 6421 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 08:06:26.065682 6421 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0220 08:06:26.065678 6421 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0220 08:06:26.065713 6421 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 08:06:26.065690 6421 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0220 08:06:26.065709 6421 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 08:06:26.065775 6421 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 08:06:26.065788 6421 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0220 08:06:26.065794 6421 factory.go:656] Stopping watch factory\\\\nI0220 08:06:26.065810 6421 handler.go:208] Removed *v1.Pod ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.608751 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.629829 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.652619 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.653038 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.653068 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.653097 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.653118 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:27Z","lastTransitionTime":"2026-02-20T08:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.652604 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.675053 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 03:55:57.781471648 +0000 UTC Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.677741 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.698557 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.722422 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.742402 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.756900 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.756952 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.756996 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.757021 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.757040 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:27Z","lastTransitionTime":"2026-02-20T08:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.860602 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.860711 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.860770 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.860801 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.860820 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:27Z","lastTransitionTime":"2026-02-20T08:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.964738 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.964825 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.964859 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.964886 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.964908 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:27Z","lastTransitionTime":"2026-02-20T08:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.002023 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql"] Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.002693 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.006535 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.006628 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.033797 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.058519 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.067907 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.068019 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.068038 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.068063 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.068083 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:28Z","lastTransitionTime":"2026-02-20T08:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.080659 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.083826 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovnkube-controller/1.log" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.086498 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhj9x\" (UniqueName: \"kubernetes.io/projected/1b664cac-f1be-4b54-b85d-fede8b4d7d31-kube-api-access-qhj9x\") pod \"ovnkube-control-plane-749d76644c-j8xql\" (UID: \"1b664cac-f1be-4b54-b85d-fede8b4d7d31\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.086574 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1b664cac-f1be-4b54-b85d-fede8b4d7d31-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-j8xql\" (UID: \"1b664cac-f1be-4b54-b85d-fede8b4d7d31\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.086850 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1b664cac-f1be-4b54-b85d-fede8b4d7d31-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-j8xql\" (UID: \"1b664cac-f1be-4b54-b85d-fede8b4d7d31\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.086956 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1b664cac-f1be-4b54-b85d-fede8b4d7d31-env-overrides\") pod \"ovnkube-control-plane-749d76644c-j8xql\" (UID: \"1b664cac-f1be-4b54-b85d-fede8b4d7d31\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.091648 4948 scope.go:117] "RemoveContainer" containerID="8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e" Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.092122 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.100427 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.123205 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.144945 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.163708 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.171727 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.171807 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.171835 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.171874 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.171898 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:28Z","lastTransitionTime":"2026-02-20T08:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.184481 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.187917 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1b664cac-f1be-4b54-b85d-fede8b4d7d31-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-j8xql\" (UID: \"1b664cac-f1be-4b54-b85d-fede8b4d7d31\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.188032 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1b664cac-f1be-4b54-b85d-fede8b4d7d31-env-overrides\") pod \"ovnkube-control-plane-749d76644c-j8xql\" (UID: \"1b664cac-f1be-4b54-b85d-fede8b4d7d31\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.188184 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhj9x\" (UniqueName: \"kubernetes.io/projected/1b664cac-f1be-4b54-b85d-fede8b4d7d31-kube-api-access-qhj9x\") pod \"ovnkube-control-plane-749d76644c-j8xql\" (UID: \"1b664cac-f1be-4b54-b85d-fede8b4d7d31\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.188236 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1b664cac-f1be-4b54-b85d-fede8b4d7d31-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-j8xql\" (UID: \"1b664cac-f1be-4b54-b85d-fede8b4d7d31\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.189491 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1b664cac-f1be-4b54-b85d-fede8b4d7d31-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-j8xql\" (UID: \"1b664cac-f1be-4b54-b85d-fede8b4d7d31\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.190179 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1b664cac-f1be-4b54-b85d-fede8b4d7d31-env-overrides\") pod \"ovnkube-control-plane-749d76644c-j8xql\" (UID: \"1b664cac-f1be-4b54-b85d-fede8b4d7d31\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.197910 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1b664cac-f1be-4b54-b85d-fede8b4d7d31-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-j8xql\" (UID: \"1b664cac-f1be-4b54-b85d-fede8b4d7d31\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.209355 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhj9x\" (UniqueName: \"kubernetes.io/projected/1b664cac-f1be-4b54-b85d-fede8b4d7d31-kube-api-access-qhj9x\") pod \"ovnkube-control-plane-749d76644c-j8xql\" (UID: \"1b664cac-f1be-4b54-b85d-fede8b4d7d31\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.222033 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.244669 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.262746 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.274796 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.274871 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.274896 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.274927 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.274953 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:28Z","lastTransitionTime":"2026-02-20T08:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.286743 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.306089 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.324898 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.343512 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3456b89afa02fb3c4b83afea773783e0ae1c65903d24789d1a742688b7745b1a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:24Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI0220 08:06:24.465835 6302 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:06:24.466938 6302 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0220 08:06:24.467064 6302 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 08:06:24.467092 6302 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0220 08:06:24.467102 6302 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0220 08:06:24.467169 6302 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 08:06:24.467179 6302 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 08:06:24.467254 6302 handler.go:208] Removed *v1.Node event handler 2\\\\nI0220 08:06:24.467267 6302 factory.go:656] Stopping watch factory\\\\nI0220 08:06:24.467295 6302 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 08:06:24.467300 6302 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0220 08:06:24.467312 6302 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 08:06:24.467326 6302 handler.go:208] Removed *v1.Node event handler 7\\\\nI0220 08:06:24.467329 6302 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:26Z\\\",\\\"message\\\":\\\"rmers/externalversions/factory.go:141\\\\nI0220 08:06:26.064417 6421 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:06:26.065536 6421 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0220 08:06:26.065606 6421 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 08:06:26.065641 6421 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 08:06:26.065650 6421 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 08:06:26.065682 6421 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0220 08:06:26.065678 6421 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0220 08:06:26.065713 6421 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 08:06:26.065690 6421 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0220 08:06:26.065709 6421 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 08:06:26.065775 6421 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 08:06:26.065788 6421 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0220 08:06:26.065794 6421 factory.go:656] Stopping watch factory\\\\nI0220 08:06:26.065810 6421 handler.go:208] Removed *v1.Pod ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: W0220 08:06:28.348509 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b664cac_f1be_4b54_b85d_fede8b4d7d31.slice/crio-d7050c82a293de4981bb1136cb34984f9983b39afd30346c219c09e99e0451c4 WatchSource:0}: Error finding container d7050c82a293de4981bb1136cb34984f9983b39afd30346c219c09e99e0451c4: Status 404 returned error can't find the container with id d7050c82a293de4981bb1136cb34984f9983b39afd30346c219c09e99e0451c4 Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.369151 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.377886 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.377933 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.377949 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.377995 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.378009 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:28Z","lastTransitionTime":"2026-02-20T08:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.389795 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.408437 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.424370 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.460314 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.477688 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.481178 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.481278 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.481305 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.481345 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.481377 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:28Z","lastTransitionTime":"2026-02-20T08:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.490441 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.490501 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.490659 4948 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.490668 4948 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.490746 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:44.490725011 +0000 UTC m=+53.465219861 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.490790 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:44.490762622 +0000 UTC m=+53.465257642 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.494543 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.519469 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.538653 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.565451 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:26Z\\\",\\\"message\\\":\\\"rmers/externalversions/factory.go:141\\\\nI0220 08:06:26.064417 6421 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:06:26.065536 6421 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0220 08:06:26.065606 6421 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 08:06:26.065641 6421 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 08:06:26.065650 6421 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 08:06:26.065682 6421 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0220 08:06:26.065678 6421 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0220 08:06:26.065713 6421 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 08:06:26.065690 6421 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0220 08:06:26.065709 6421 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 08:06:26.065775 6421 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 08:06:26.065788 6421 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0220 08:06:26.065794 6421 factory.go:656] Stopping watch factory\\\\nI0220 08:06:26.065810 6421 handler.go:208] Removed *v1.Pod ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.585492 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.585571 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.585591 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.585618 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.585638 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:28Z","lastTransitionTime":"2026-02-20T08:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.587135 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.591321 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.591485 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:06:44.591456824 +0000 UTC m=+53.565951654 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.591532 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.591621 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.591740 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.591758 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.591774 4948 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.591825 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:44.591814382 +0000 UTC m=+53.566309212 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.591883 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.591919 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.591943 4948 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.592085 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:44.592050468 +0000 UTC m=+53.566545348 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.608418 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.625147 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.651792 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.676303 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 19:08:58.852604819 +0000 UTC Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.677336 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.693579 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.693639 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.693727 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.693790 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.693811 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:28Z","lastTransitionTime":"2026-02-20T08:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.705024 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.721496 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.721539 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.721623 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.721671 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.721804 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.722003 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.722092 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.740836 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.770104 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-xd86g"] Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.770868 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.771007 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.793581 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs\") pod \"network-metrics-daemon-xd86g\" (UID: \"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\") " pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.793673 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdbl2\" (UniqueName: \"kubernetes.io/projected/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-kube-api-access-gdbl2\") pod \"network-metrics-daemon-xd86g\" (UID: \"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\") " pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.797374 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.797426 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.797440 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.797461 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.797476 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:28Z","lastTransitionTime":"2026-02-20T08:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.821055 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:26Z\\\",\\\"message\\\":\\\"rmers/externalversions/factory.go:141\\\\nI0220 08:06:26.064417 6421 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:06:26.065536 6421 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0220 08:06:26.065606 6421 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 08:06:26.065641 6421 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 08:06:26.065650 6421 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 08:06:26.065682 6421 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0220 08:06:26.065678 6421 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0220 08:06:26.065713 6421 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 08:06:26.065690 6421 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0220 08:06:26.065709 6421 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 08:06:26.065775 6421 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 08:06:26.065788 6421 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0220 08:06:26.065794 6421 factory.go:656] Stopping watch factory\\\\nI0220 08:06:26.065810 6421 handler.go:208] Removed *v1.Pod ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.843020 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.850176 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.850218 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.850228 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.850247 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.850259 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:28Z","lastTransitionTime":"2026-02-20T08:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.858174 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.866174 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.870019 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.870058 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.870069 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.870086 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.870117 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:28Z","lastTransitionTime":"2026-02-20T08:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.875243 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.882842 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.886512 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.887772 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.887825 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.887834 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.887852 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.887863 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:28Z","lastTransitionTime":"2026-02-20T08:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.894325 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs\") pod \"network-metrics-daemon-xd86g\" (UID: \"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\") " pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.894362 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdbl2\" (UniqueName: \"kubernetes.io/projected/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-kube-api-access-gdbl2\") pod \"network-metrics-daemon-xd86g\" (UID: \"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\") " pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.894474 4948 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.894530 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs podName:62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd nodeName:}" failed. No retries permitted until 2026-02-20 08:06:29.394516213 +0000 UTC m=+38.369011033 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs") pod "network-metrics-daemon-xd86g" (UID: "62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.900550 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.902464 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.905856 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.905887 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.905894 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.905909 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.905919 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:28Z","lastTransitionTime":"2026-02-20T08:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.915910 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdbl2\" (UniqueName: \"kubernetes.io/projected/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-kube-api-access-gdbl2\") pod \"network-metrics-daemon-xd86g\" (UID: \"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\") " pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.917716 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.920113 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.924666 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.924718 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.924731 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.924751 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.924764 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:28Z","lastTransitionTime":"2026-02-20T08:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.932456 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.938471 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.938575 4948 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.940139 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.940165 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.940176 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.940191 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.940202 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:28Z","lastTransitionTime":"2026-02-20T08:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.947747 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.968117 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.986896 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.005385 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.030886 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.042733 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.042787 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.042800 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.042818 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.042831 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:29Z","lastTransitionTime":"2026-02-20T08:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.047738 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.061351 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.075788 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.097596 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xd86g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xd86g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.100732 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" event={"ID":"1b664cac-f1be-4b54-b85d-fede8b4d7d31","Type":"ContainerStarted","Data":"b5cc8878e457802d4970c4db52fc332cf96d26ab80c90cbe1c3160b8c02b385e"} Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.100854 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" event={"ID":"1b664cac-f1be-4b54-b85d-fede8b4d7d31","Type":"ContainerStarted","Data":"c6cfa24bc2ee3daae226cdfb1ec830d8a97c86eed37ca8ce2d0c10878627bcc4"} Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.100881 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" event={"ID":"1b664cac-f1be-4b54-b85d-fede8b4d7d31","Type":"ContainerStarted","Data":"d7050c82a293de4981bb1136cb34984f9983b39afd30346c219c09e99e0451c4"} Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.125069 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.145712 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.146451 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.146501 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.146520 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.146547 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.146599 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:29Z","lastTransitionTime":"2026-02-20T08:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.169194 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.190188 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.226662 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:26Z\\\",\\\"message\\\":\\\"rmers/externalversions/factory.go:141\\\\nI0220 08:06:26.064417 6421 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:06:26.065536 6421 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0220 08:06:26.065606 6421 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 08:06:26.065641 6421 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 08:06:26.065650 6421 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 08:06:26.065682 6421 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0220 08:06:26.065678 6421 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0220 08:06:26.065713 6421 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 08:06:26.065690 6421 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0220 08:06:26.065709 6421 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 08:06:26.065775 6421 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 08:06:26.065788 6421 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0220 08:06:26.065794 6421 factory.go:656] Stopping watch factory\\\\nI0220 08:06:26.065810 6421 handler.go:208] Removed *v1.Pod ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.249968 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.250105 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.250131 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.250169 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.250196 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:29Z","lastTransitionTime":"2026-02-20T08:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.254155 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.276806 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.297824 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.316968 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.338033 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.353486 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.353557 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.353574 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.353601 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.353619 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:29Z","lastTransitionTime":"2026-02-20T08:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.356321 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.377638 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.400554 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs\") pod \"network-metrics-daemon-xd86g\" (UID: \"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\") " pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:29 crc kubenswrapper[4948]: E0220 08:06:29.400826 4948 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.400763 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: E0220 08:06:29.400930 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs podName:62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd nodeName:}" failed. No retries permitted until 2026-02-20 08:06:30.400901223 +0000 UTC m=+39.375396083 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs") pod "network-metrics-daemon-xd86g" (UID: "62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.425835 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.444523 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.456687 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.456767 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.456792 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.456827 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.456854 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:29Z","lastTransitionTime":"2026-02-20T08:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.462719 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6cfa24bc2ee3daae226cdfb1ec830d8a97c86eed37ca8ce2d0c10878627bcc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cc8878e457802d4970c4db52fc332cf96d26ab80c90cbe1c3160b8c02b385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.483847 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xd86g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xd86g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.559899 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.559960 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.560256 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.560291 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.560310 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:29Z","lastTransitionTime":"2026-02-20T08:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.663185 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.663259 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.663283 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.663314 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.663342 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:29Z","lastTransitionTime":"2026-02-20T08:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.676692 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 18:37:38.369965687 +0000 UTC Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.766639 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.766702 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.766722 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.766749 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.766771 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:29Z","lastTransitionTime":"2026-02-20T08:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.870708 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.870766 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.870788 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.870813 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.870832 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:29Z","lastTransitionTime":"2026-02-20T08:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.974570 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.974635 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.974657 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.974684 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.974703 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:29Z","lastTransitionTime":"2026-02-20T08:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.078363 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.078428 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.078445 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.078470 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.078487 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:30Z","lastTransitionTime":"2026-02-20T08:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.182199 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.182280 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.182293 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.182320 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.182335 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:30Z","lastTransitionTime":"2026-02-20T08:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.285678 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.286031 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.286168 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.286310 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.286445 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:30Z","lastTransitionTime":"2026-02-20T08:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.389518 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.389583 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.389604 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.389631 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.389652 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:30Z","lastTransitionTime":"2026-02-20T08:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.413608 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs\") pod \"network-metrics-daemon-xd86g\" (UID: \"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\") " pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:30 crc kubenswrapper[4948]: E0220 08:06:30.413784 4948 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 08:06:30 crc kubenswrapper[4948]: E0220 08:06:30.414127 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs podName:62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd nodeName:}" failed. No retries permitted until 2026-02-20 08:06:32.414096922 +0000 UTC m=+41.388591782 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs") pod "network-metrics-daemon-xd86g" (UID: "62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.493052 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.493471 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.493605 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.493761 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.493894 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:30Z","lastTransitionTime":"2026-02-20T08:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.598121 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.598192 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.598210 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.598238 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.598258 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:30Z","lastTransitionTime":"2026-02-20T08:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.677707 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 21:13:07.062013621 +0000 UTC Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.701321 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.701391 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.701412 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.701439 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.701462 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:30Z","lastTransitionTime":"2026-02-20T08:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.722425 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.722484 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:30 crc kubenswrapper[4948]: E0220 08:06:30.722647 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.722673 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.722715 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:30 crc kubenswrapper[4948]: E0220 08:06:30.722851 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:30 crc kubenswrapper[4948]: E0220 08:06:30.723013 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:30 crc kubenswrapper[4948]: E0220 08:06:30.723136 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.804752 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.804811 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.804830 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.804857 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.804876 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:30Z","lastTransitionTime":"2026-02-20T08:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.908912 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.909017 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.909040 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.909072 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.909092 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:30Z","lastTransitionTime":"2026-02-20T08:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.013218 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.013313 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.013335 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.013361 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.013380 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:31Z","lastTransitionTime":"2026-02-20T08:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.116200 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.116267 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.116285 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.116314 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.116335 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:31Z","lastTransitionTime":"2026-02-20T08:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.220225 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.220308 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.220328 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.220364 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.220389 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:31Z","lastTransitionTime":"2026-02-20T08:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.324128 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.324199 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.324219 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.324247 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.324269 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:31Z","lastTransitionTime":"2026-02-20T08:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.428019 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.428079 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.428091 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.428113 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.428130 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:31Z","lastTransitionTime":"2026-02-20T08:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.531580 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.531681 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.531700 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.531728 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.531746 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:31Z","lastTransitionTime":"2026-02-20T08:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.635190 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.635253 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.635271 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.635298 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.635318 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:31Z","lastTransitionTime":"2026-02-20T08:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.678792 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 15:35:05.395643427 +0000 UTC Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.738320 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.738394 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.738413 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.738441 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.738460 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:31Z","lastTransitionTime":"2026-02-20T08:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.741529 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:31Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.774070 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:26Z\\\",\\\"message\\\":\\\"rmers/externalversions/factory.go:141\\\\nI0220 08:06:26.064417 6421 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:06:26.065536 6421 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0220 08:06:26.065606 6421 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 08:06:26.065641 6421 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 08:06:26.065650 6421 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 08:06:26.065682 6421 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0220 08:06:26.065678 6421 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0220 08:06:26.065713 6421 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 08:06:26.065690 6421 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0220 08:06:26.065709 6421 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 08:06:26.065775 6421 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 08:06:26.065788 6421 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0220 08:06:26.065794 6421 factory.go:656] Stopping watch factory\\\\nI0220 08:06:26.065810 6421 handler.go:208] Removed *v1.Pod ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:31Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.796335 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:31Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.815093 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:31Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.840823 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.840877 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.840889 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.840910 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.840935 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:31Z","lastTransitionTime":"2026-02-20T08:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.848040 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:31Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.873277 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:31Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.899259 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:31Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.920451 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:31Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.944128 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:31Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.944475 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.944541 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.944558 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.944584 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.944601 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:31Z","lastTransitionTime":"2026-02-20T08:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.963114 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:31Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.981468 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:31Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.998240 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:31Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.012953 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xd86g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xd86g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:32Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.041428 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:32Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.047451 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.047531 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.047551 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.047582 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.047602 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:32Z","lastTransitionTime":"2026-02-20T08:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.064242 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:32Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.080689 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:32Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.100862 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6cfa24bc2ee3daae226cdfb1ec830d8a97c86eed37ca8ce2d0c10878627bcc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cc8878e457802d4970c4db52fc332cf96d26ab80c90cbe1c3160b8c02b385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:32Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.151648 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.151711 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.151728 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.151753 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.151770 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:32Z","lastTransitionTime":"2026-02-20T08:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.255726 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.255784 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.255803 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.255832 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.255850 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:32Z","lastTransitionTime":"2026-02-20T08:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.359668 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.360087 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.360265 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.360321 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.360343 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:32Z","lastTransitionTime":"2026-02-20T08:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.438658 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs\") pod \"network-metrics-daemon-xd86g\" (UID: \"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\") " pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:32 crc kubenswrapper[4948]: E0220 08:06:32.438855 4948 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 08:06:32 crc kubenswrapper[4948]: E0220 08:06:32.439000 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs podName:62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd nodeName:}" failed. No retries permitted until 2026-02-20 08:06:36.438943386 +0000 UTC m=+45.413438246 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs") pod "network-metrics-daemon-xd86g" (UID: "62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.463386 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.463469 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.463492 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.463525 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.463550 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:32Z","lastTransitionTime":"2026-02-20T08:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.566692 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.566748 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.566765 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.566794 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.566812 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:32Z","lastTransitionTime":"2026-02-20T08:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.670301 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.670372 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.670401 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.670438 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.670465 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:32Z","lastTransitionTime":"2026-02-20T08:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.679253 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 17:06:39.819894113 +0000 UTC Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.721642 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.721669 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.721677 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.721841 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:32 crc kubenswrapper[4948]: E0220 08:06:32.722063 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:32 crc kubenswrapper[4948]: E0220 08:06:32.722237 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:32 crc kubenswrapper[4948]: E0220 08:06:32.722373 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:32 crc kubenswrapper[4948]: E0220 08:06:32.722523 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.732549 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.733782 4948 scope.go:117] "RemoveContainer" containerID="8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e" Feb 20 08:06:32 crc kubenswrapper[4948]: E0220 08:06:32.734072 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.773446 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.773509 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.773527 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.773551 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.773571 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:32Z","lastTransitionTime":"2026-02-20T08:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.876934 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.877030 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.877052 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.877124 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.877144 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:32Z","lastTransitionTime":"2026-02-20T08:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.980423 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.980485 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.980501 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.980525 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.980543 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:32Z","lastTransitionTime":"2026-02-20T08:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.084226 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.084313 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.084340 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.084371 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.084393 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:33Z","lastTransitionTime":"2026-02-20T08:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.187824 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.187888 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.187912 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.187941 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.187963 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:33Z","lastTransitionTime":"2026-02-20T08:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.291784 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.291835 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.291850 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.291873 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.291889 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:33Z","lastTransitionTime":"2026-02-20T08:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.401373 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.401436 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.401454 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.401484 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.401502 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:33Z","lastTransitionTime":"2026-02-20T08:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.505029 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.505106 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.505123 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.505151 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.505170 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:33Z","lastTransitionTime":"2026-02-20T08:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.608886 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.608954 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.609009 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.609039 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.609061 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:33Z","lastTransitionTime":"2026-02-20T08:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.680127 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 13:44:21.15647115 +0000 UTC Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.711883 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.711961 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.712012 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.712041 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.712060 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:33Z","lastTransitionTime":"2026-02-20T08:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.815706 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.815798 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.815818 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.815844 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.815864 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:33Z","lastTransitionTime":"2026-02-20T08:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.919213 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.919274 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.919292 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.919320 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.919348 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:33Z","lastTransitionTime":"2026-02-20T08:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.022318 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.022378 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.022396 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.022423 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.022445 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:34Z","lastTransitionTime":"2026-02-20T08:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.124767 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.124881 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.124907 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.124937 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.124958 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:34Z","lastTransitionTime":"2026-02-20T08:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.229231 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.229318 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.229347 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.229381 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.229406 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:34Z","lastTransitionTime":"2026-02-20T08:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.332732 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.332777 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.332794 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.332814 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.332832 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:34Z","lastTransitionTime":"2026-02-20T08:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.436424 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.436513 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.436529 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.436553 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.436569 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:34Z","lastTransitionTime":"2026-02-20T08:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.540231 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.540293 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.540310 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.540337 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.540357 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:34Z","lastTransitionTime":"2026-02-20T08:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.643764 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.643844 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.643867 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.643895 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.643919 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:34Z","lastTransitionTime":"2026-02-20T08:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.681136 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 15:02:47.403721519 +0000 UTC Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.721478 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.721543 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.721499 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.721479 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:34 crc kubenswrapper[4948]: E0220 08:06:34.721744 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:06:34 crc kubenswrapper[4948]: E0220 08:06:34.721895 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:34 crc kubenswrapper[4948]: E0220 08:06:34.722106 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:34 crc kubenswrapper[4948]: E0220 08:06:34.722308 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.746847 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.746917 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.746943 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.747011 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.747062 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:34Z","lastTransitionTime":"2026-02-20T08:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.850541 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.850597 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.850610 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.850631 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.850644 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:34Z","lastTransitionTime":"2026-02-20T08:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.953265 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.953329 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.953347 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.953374 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.953393 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:34Z","lastTransitionTime":"2026-02-20T08:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.056809 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.056884 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.056906 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.056938 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.056958 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:35Z","lastTransitionTime":"2026-02-20T08:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.159516 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.159641 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.159666 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.159704 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.159723 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:35Z","lastTransitionTime":"2026-02-20T08:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.263243 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.263317 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.263336 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.263363 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.263384 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:35Z","lastTransitionTime":"2026-02-20T08:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.365726 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.365773 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.365785 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.365803 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.365813 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:35Z","lastTransitionTime":"2026-02-20T08:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.468931 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.469077 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.469092 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.469112 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.469124 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:35Z","lastTransitionTime":"2026-02-20T08:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.571944 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.571997 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.572008 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.572026 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.572038 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:35Z","lastTransitionTime":"2026-02-20T08:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.675297 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.675344 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.675356 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.675381 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.675400 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:35Z","lastTransitionTime":"2026-02-20T08:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.681984 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 03:58:01.055675842 +0000 UTC Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.778040 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.778107 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.778132 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.778159 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.778179 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:35Z","lastTransitionTime":"2026-02-20T08:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.881865 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.881924 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.881937 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.881960 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.882001 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:35Z","lastTransitionTime":"2026-02-20T08:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.984640 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.984864 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.984886 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.984913 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.984931 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:35Z","lastTransitionTime":"2026-02-20T08:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.088019 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.088089 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.088106 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.088134 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.088155 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:36Z","lastTransitionTime":"2026-02-20T08:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.191886 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.191953 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.192012 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.192047 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.192071 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:36Z","lastTransitionTime":"2026-02-20T08:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.295220 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.295319 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.295340 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.295363 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.295383 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:36Z","lastTransitionTime":"2026-02-20T08:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.398275 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.398335 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.398352 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.398381 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.398400 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:36Z","lastTransitionTime":"2026-02-20T08:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.495565 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs\") pod \"network-metrics-daemon-xd86g\" (UID: \"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\") " pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:36 crc kubenswrapper[4948]: E0220 08:06:36.495754 4948 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 08:06:36 crc kubenswrapper[4948]: E0220 08:06:36.495812 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs podName:62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd nodeName:}" failed. No retries permitted until 2026-02-20 08:06:44.495796244 +0000 UTC m=+53.470291064 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs") pod "network-metrics-daemon-xd86g" (UID: "62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.501594 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.501631 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.501643 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.501662 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.501676 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:36Z","lastTransitionTime":"2026-02-20T08:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.604998 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.605066 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.605090 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.605118 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.605141 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:36Z","lastTransitionTime":"2026-02-20T08:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.683121 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 08:58:33.057202881 +0000 UTC Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.708054 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.708117 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.708138 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.708163 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.708181 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:36Z","lastTransitionTime":"2026-02-20T08:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.721597 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.721609 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:36 crc kubenswrapper[4948]: E0220 08:06:36.721755 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.721632 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.721631 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:36 crc kubenswrapper[4948]: E0220 08:06:36.721926 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:06:36 crc kubenswrapper[4948]: E0220 08:06:36.722120 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:36 crc kubenswrapper[4948]: E0220 08:06:36.722267 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.811000 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.811062 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.811078 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.811102 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.811120 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:36Z","lastTransitionTime":"2026-02-20T08:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.914138 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.914204 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.914230 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.914263 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.914289 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:36Z","lastTransitionTime":"2026-02-20T08:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.016942 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.017098 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.017121 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.017151 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.017188 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:37Z","lastTransitionTime":"2026-02-20T08:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.119891 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.119960 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.120016 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.120051 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.120074 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:37Z","lastTransitionTime":"2026-02-20T08:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.222922 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.223007 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.223028 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.223053 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.223072 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:37Z","lastTransitionTime":"2026-02-20T08:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.326833 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.326897 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.326914 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.326940 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.326958 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:37Z","lastTransitionTime":"2026-02-20T08:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.429727 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.429767 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.429778 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.429795 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.429807 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:37Z","lastTransitionTime":"2026-02-20T08:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.532648 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.532718 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.532737 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.532763 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.532781 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:37Z","lastTransitionTime":"2026-02-20T08:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.636587 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.636636 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.636645 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.636660 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.636671 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:37Z","lastTransitionTime":"2026-02-20T08:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.683312 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 03:19:44.299217979 +0000 UTC Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.739861 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.739937 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.739962 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.740049 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.740078 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:37Z","lastTransitionTime":"2026-02-20T08:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.843179 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.843235 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.843253 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.843276 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.843292 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:37Z","lastTransitionTime":"2026-02-20T08:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.945920 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.946018 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.946030 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.946056 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.946072 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:37Z","lastTransitionTime":"2026-02-20T08:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.049043 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.049128 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.049143 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.049170 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.049186 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:38Z","lastTransitionTime":"2026-02-20T08:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.151458 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.151830 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.152057 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.152237 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.152391 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:38Z","lastTransitionTime":"2026-02-20T08:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.255696 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.255747 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.255764 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.255783 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.255794 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:38Z","lastTransitionTime":"2026-02-20T08:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.359497 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.359557 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.359570 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.359592 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.359607 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:38Z","lastTransitionTime":"2026-02-20T08:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.462697 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.462750 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.462765 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.462785 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.462799 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:38Z","lastTransitionTime":"2026-02-20T08:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.566144 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.566229 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.566267 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.566304 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.566330 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:38Z","lastTransitionTime":"2026-02-20T08:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.669578 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.669669 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.669693 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.669728 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.669752 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:38Z","lastTransitionTime":"2026-02-20T08:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.683842 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 21:33:02.940916729 +0000 UTC Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.722252 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.722369 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.722431 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:38 crc kubenswrapper[4948]: E0220 08:06:38.722479 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.722390 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:38 crc kubenswrapper[4948]: E0220 08:06:38.722582 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:38 crc kubenswrapper[4948]: E0220 08:06:38.722800 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:38 crc kubenswrapper[4948]: E0220 08:06:38.722932 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.772962 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.773071 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.773102 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.773132 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.773154 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:38Z","lastTransitionTime":"2026-02-20T08:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.876800 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.876843 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.876854 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.876871 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.876885 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:38Z","lastTransitionTime":"2026-02-20T08:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.980010 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.980097 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.980160 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.980194 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.980217 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:38Z","lastTransitionTime":"2026-02-20T08:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.057735 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.057791 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.057803 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.057823 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.057837 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:39Z","lastTransitionTime":"2026-02-20T08:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:39 crc kubenswrapper[4948]: E0220 08:06:39.080043 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:39Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.085533 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.085593 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.085626 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.085653 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.085671 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:39Z","lastTransitionTime":"2026-02-20T08:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:39 crc kubenswrapper[4948]: E0220 08:06:39.106520 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:39Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.111501 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.111558 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.111576 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.111600 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.111621 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:39Z","lastTransitionTime":"2026-02-20T08:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:39 crc kubenswrapper[4948]: E0220 08:06:39.131269 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:39Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.137022 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.137175 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.137288 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.137340 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.137403 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:39Z","lastTransitionTime":"2026-02-20T08:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:39 crc kubenswrapper[4948]: E0220 08:06:39.160211 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:39Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.166825 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.166887 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.166907 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.166933 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.166952 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:39Z","lastTransitionTime":"2026-02-20T08:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:39 crc kubenswrapper[4948]: E0220 08:06:39.188484 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:39Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:39 crc kubenswrapper[4948]: E0220 08:06:39.188719 4948 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.191094 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.191149 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.191161 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.191182 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.191198 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:39Z","lastTransitionTime":"2026-02-20T08:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.295116 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.295701 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.296032 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.296242 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.296449 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:39Z","lastTransitionTime":"2026-02-20T08:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.400714 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.400773 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.400794 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.400825 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.400849 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:39Z","lastTransitionTime":"2026-02-20T08:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.504030 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.504092 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.504116 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.504148 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.504175 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:39Z","lastTransitionTime":"2026-02-20T08:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.606923 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.607015 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.607033 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.607060 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.607079 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:39Z","lastTransitionTime":"2026-02-20T08:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.684913 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 11:36:56.624396106 +0000 UTC Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.709430 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.709500 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.709528 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.709558 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.709578 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:39Z","lastTransitionTime":"2026-02-20T08:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.813324 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.813408 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.813427 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.813454 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.813475 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:39Z","lastTransitionTime":"2026-02-20T08:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.916956 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.917060 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.917084 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.917118 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.917143 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:39Z","lastTransitionTime":"2026-02-20T08:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.020524 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.020591 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.020610 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.020638 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.020658 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:40Z","lastTransitionTime":"2026-02-20T08:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.124461 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.124533 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.124551 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.124577 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.124598 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:40Z","lastTransitionTime":"2026-02-20T08:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.233666 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.233751 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.233775 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.233810 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.233835 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:40Z","lastTransitionTime":"2026-02-20T08:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.336597 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.336650 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.336661 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.336680 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.336693 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:40Z","lastTransitionTime":"2026-02-20T08:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.439026 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.439094 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.439117 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.439146 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.439168 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:40Z","lastTransitionTime":"2026-02-20T08:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.543623 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.543707 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.543732 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.543767 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.543790 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:40Z","lastTransitionTime":"2026-02-20T08:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.648099 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.648176 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.648200 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.648231 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.648253 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:40Z","lastTransitionTime":"2026-02-20T08:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.685625 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 03:29:19.944070432 +0000 UTC Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.722357 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.722399 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.722554 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:40 crc kubenswrapper[4948]: E0220 08:06:40.722546 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.722698 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:40 crc kubenswrapper[4948]: E0220 08:06:40.722836 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:06:40 crc kubenswrapper[4948]: E0220 08:06:40.723071 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:40 crc kubenswrapper[4948]: E0220 08:06:40.723203 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.751628 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.752053 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.752245 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.752439 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.752636 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:40Z","lastTransitionTime":"2026-02-20T08:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.856797 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.856878 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.856901 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.856929 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.856950 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:40Z","lastTransitionTime":"2026-02-20T08:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.960562 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.960661 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.960685 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.960719 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.960743 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:40Z","lastTransitionTime":"2026-02-20T08:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.064301 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.064367 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.064384 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.064409 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.064437 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:41Z","lastTransitionTime":"2026-02-20T08:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.168094 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.168183 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.168202 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.168226 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.168246 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:41Z","lastTransitionTime":"2026-02-20T08:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.271488 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.271783 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.271933 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.272138 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.272283 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:41Z","lastTransitionTime":"2026-02-20T08:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.375474 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.375542 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.375561 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.375590 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.375615 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:41Z","lastTransitionTime":"2026-02-20T08:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.478616 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.478691 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.478709 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.478736 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.478756 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:41Z","lastTransitionTime":"2026-02-20T08:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.581862 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.581951 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.582005 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.582032 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.582051 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:41Z","lastTransitionTime":"2026-02-20T08:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.685253 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.685331 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.685350 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.685382 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.685403 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:41Z","lastTransitionTime":"2026-02-20T08:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.686076 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 23:42:08.028612677 +0000 UTC Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.745814 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:41Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.763931 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:41Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.781706 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6cfa24bc2ee3daae226cdfb1ec830d8a97c86eed37ca8ce2d0c10878627bcc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cc8878e457802d4970c4db52fc332cf96d26ab80c90cbe1c3160b8c02b385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:41Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.787861 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.787935 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.787954 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.788013 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.788035 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:41Z","lastTransitionTime":"2026-02-20T08:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.798465 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xd86g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xd86g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:41Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.840191 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:41Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.866581 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:41Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.884541 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:41Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.890796 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.890871 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.890894 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.890948 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.891003 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:41Z","lastTransitionTime":"2026-02-20T08:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.911467 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:41Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.932666 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:41Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.994637 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.994764 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.994787 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.994823 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.994855 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:41Z","lastTransitionTime":"2026-02-20T08:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.012248 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:26Z\\\",\\\"message\\\":\\\"rmers/externalversions/factory.go:141\\\\nI0220 08:06:26.064417 6421 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:06:26.065536 6421 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0220 08:06:26.065606 6421 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 08:06:26.065641 6421 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 08:06:26.065650 6421 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 08:06:26.065682 6421 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0220 08:06:26.065678 6421 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0220 08:06:26.065713 6421 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 08:06:26.065690 6421 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0220 08:06:26.065709 6421 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 08:06:26.065775 6421 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 08:06:26.065788 6421 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0220 08:06:26.065794 6421 factory.go:656] Stopping watch factory\\\\nI0220 08:06:26.065810 6421 handler.go:208] Removed *v1.Pod ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:42Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.042787 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:42Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.055021 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:42Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.076321 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:42Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.088102 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:42Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.098418 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.098470 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.098482 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.098502 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.098516 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:42Z","lastTransitionTime":"2026-02-20T08:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.102636 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:42Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.114881 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:42Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.128603 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:42Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.202346 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.202415 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.202432 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.202457 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.202476 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:42Z","lastTransitionTime":"2026-02-20T08:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.305331 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.305393 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.305405 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.305425 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.305745 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:42Z","lastTransitionTime":"2026-02-20T08:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.408648 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.408710 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.408728 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.408757 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.408778 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:42Z","lastTransitionTime":"2026-02-20T08:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.512282 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.512364 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.512387 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.512414 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.512441 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:42Z","lastTransitionTime":"2026-02-20T08:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.615510 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.615596 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.615633 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.615667 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.615698 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:42Z","lastTransitionTime":"2026-02-20T08:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.687105 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 21:55:29.672665013 +0000 UTC Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.718900 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.718967 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.719022 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.719051 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.719071 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:42Z","lastTransitionTime":"2026-02-20T08:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.721727 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.721822 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.721842 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.721727 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:42 crc kubenswrapper[4948]: E0220 08:06:42.721936 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:06:42 crc kubenswrapper[4948]: E0220 08:06:42.722118 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:42 crc kubenswrapper[4948]: E0220 08:06:42.722291 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:42 crc kubenswrapper[4948]: E0220 08:06:42.722496 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.822435 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.822511 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.822530 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.822557 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.822578 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:42Z","lastTransitionTime":"2026-02-20T08:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.925467 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.925557 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.925591 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.925620 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.925638 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:42Z","lastTransitionTime":"2026-02-20T08:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.028770 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.028827 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.028880 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.028911 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.029519 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:43Z","lastTransitionTime":"2026-02-20T08:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.153907 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.153967 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.154027 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.154064 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.154087 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:43Z","lastTransitionTime":"2026-02-20T08:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.258201 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.258286 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.258307 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.258344 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.258367 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:43Z","lastTransitionTime":"2026-02-20T08:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.362384 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.362484 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.362503 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.362529 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.362549 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:43Z","lastTransitionTime":"2026-02-20T08:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.465661 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.465731 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.465751 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.465783 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.465811 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:43Z","lastTransitionTime":"2026-02-20T08:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.568541 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.568651 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.568671 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.568699 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.568718 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:43Z","lastTransitionTime":"2026-02-20T08:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.671227 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.671374 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.671399 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.671422 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.671443 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:43Z","lastTransitionTime":"2026-02-20T08:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.688157 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 16:28:12.724100466 +0000 UTC Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.723382 4948 scope.go:117] "RemoveContainer" containerID="8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.779536 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.779599 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.779625 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.779656 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.779676 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:43Z","lastTransitionTime":"2026-02-20T08:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.883083 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.884155 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.884210 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.884248 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.884274 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:43Z","lastTransitionTime":"2026-02-20T08:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:43.991472 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:43.991524 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:43.991542 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:43.991565 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:43.991583 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:43Z","lastTransitionTime":"2026-02-20T08:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.095909 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.095962 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.096021 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.096050 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.096071 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:44Z","lastTransitionTime":"2026-02-20T08:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.161848 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovnkube-controller/1.log" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.166817 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerStarted","Data":"bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf"} Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.167535 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.196235 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.198756 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.198809 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.198835 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.198869 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.198894 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:44Z","lastTransitionTime":"2026-02-20T08:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.216466 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.241752 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.265492 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.292837 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.302162 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.302213 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.302230 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.302249 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.302261 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:44Z","lastTransitionTime":"2026-02-20T08:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.313284 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.339388 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6cfa24bc2ee3daae226cdfb1ec830d8a97c86eed37ca8ce2d0c10878627bcc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cc8878e457802d4970c4db52fc332cf96d26ab80c90cbe1c3160b8c02b385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.364078 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xd86g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xd86g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.405473 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.405572 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.405591 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.405613 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.405626 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:44Z","lastTransitionTime":"2026-02-20T08:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.406306 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.433721 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.454460 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.474657 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.491616 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.491937 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.491790 4948 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.491590 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.492105 4948 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.492347 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 08:07:16.492324771 +0000 UTC m=+85.466819601 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.492496 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 08:07:16.492432243 +0000 UTC m=+85.466927073 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.508517 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.508580 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.508595 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.508618 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.508636 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:44Z","lastTransitionTime":"2026-02-20T08:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.518450 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:26Z\\\",\\\"message\\\":\\\"rmers/externalversions/factory.go:141\\\\nI0220 08:06:26.064417 6421 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:06:26.065536 6421 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0220 08:06:26.065606 6421 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 08:06:26.065641 6421 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 08:06:26.065650 6421 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 08:06:26.065682 6421 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0220 08:06:26.065678 6421 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0220 08:06:26.065713 6421 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 08:06:26.065690 6421 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0220 08:06:26.065709 6421 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 08:06:26.065775 6421 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 08:06:26.065788 6421 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0220 08:06:26.065794 6421 factory.go:656] Stopping watch factory\\\\nI0220 08:06:26.065810 6421 handler.go:208] Removed *v1.Pod ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.533774 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.550204 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.568902 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.593367 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.593653 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:07:16.593604247 +0000 UTC m=+85.568099107 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.593766 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.593911 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs\") pod \"network-metrics-daemon-xd86g\" (UID: \"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\") " pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.593951 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.593999 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.594016 4948 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.594041 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.594078 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-20 08:07:16.594057087 +0000 UTC m=+85.568551917 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.594163 4948 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.594187 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.594315 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.594335 4948 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.594319 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs podName:62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd nodeName:}" failed. No retries permitted until 2026-02-20 08:07:00.594274323 +0000 UTC m=+69.568769333 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs") pod "network-metrics-daemon-xd86g" (UID: "62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.594395 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-20 08:07:16.594384945 +0000 UTC m=+85.568879765 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.611406 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.611442 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.611453 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.611484 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.611503 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:44Z","lastTransitionTime":"2026-02-20T08:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.688924 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 16:49:01.154887844 +0000 UTC Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.714445 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.714491 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.714500 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.714518 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.714530 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:44Z","lastTransitionTime":"2026-02-20T08:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.721794 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.721830 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.721863 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.721909 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.722049 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.721794 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.722120 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.722169 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.818120 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.818183 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.818199 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.818228 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.818248 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:44Z","lastTransitionTime":"2026-02-20T08:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.920917 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.921076 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.921105 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.921142 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.921173 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:44Z","lastTransitionTime":"2026-02-20T08:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.025379 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.025437 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.025449 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.025469 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.025481 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:45Z","lastTransitionTime":"2026-02-20T08:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.129213 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.129570 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.129700 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.129894 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.130056 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:45Z","lastTransitionTime":"2026-02-20T08:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.175530 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovnkube-controller/2.log" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.177813 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovnkube-controller/1.log" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.186209 4948 generic.go:334] "Generic (PLEG): container finished" podID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerID="bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf" exitCode=1 Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.186336 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerDied","Data":"bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf"} Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.186692 4948 scope.go:117] "RemoveContainer" containerID="8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.187693 4948 scope.go:117] "RemoveContainer" containerID="bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf" Feb 20 08:06:45 crc kubenswrapper[4948]: E0220 08:06:45.188035 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.208740 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6cfa24bc2ee3daae226cdfb1ec830d8a97c86eed37ca8ce2d0c10878627bcc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cc8878e457802d4970c4db52fc332cf96d26ab80c90cbe1c3160b8c02b385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.229451 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xd86g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xd86g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.234551 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.234613 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.234630 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.234657 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.234682 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:45Z","lastTransitionTime":"2026-02-20T08:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.274057 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.294893 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.316378 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.334564 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.338552 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.338629 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.338657 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.338681 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.338703 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:45Z","lastTransitionTime":"2026-02-20T08:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.346455 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.350266 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.367760 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.400509 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:26Z\\\",\\\"message\\\":\\\"rmers/externalversions/factory.go:141\\\\nI0220 08:06:26.064417 6421 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:06:26.065536 6421 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0220 08:06:26.065606 6421 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 08:06:26.065641 6421 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 08:06:26.065650 6421 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 08:06:26.065682 6421 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0220 08:06:26.065678 6421 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0220 08:06:26.065713 6421 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 08:06:26.065690 6421 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0220 08:06:26.065709 6421 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 08:06:26.065775 6421 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 08:06:26.065788 6421 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0220 08:06:26.065794 6421 factory.go:656] Stopping watch factory\\\\nI0220 08:06:26.065810 6421 handler.go:208] Removed *v1.Pod ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:44Z\\\",\\\"message\\\":\\\"ate admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z]\\\\nI0220 08:06:44.892233 6647 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/community-operators]} name:Service_openshift-marketplace/community-operators_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.189:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d389393c-7ba9-422c-b3f5-06e391d537d2}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.423180 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.442233 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.442355 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.442384 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.442416 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.442447 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:45Z","lastTransitionTime":"2026-02-20T08:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.444015 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.465008 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.490351 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.513940 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.536153 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.545457 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.545544 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.545568 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.545599 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.545623 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:45Z","lastTransitionTime":"2026-02-20T08:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.559546 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.581310 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.600516 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.621056 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.642052 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.648465 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.648511 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.648522 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.648538 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.648551 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:45Z","lastTransitionTime":"2026-02-20T08:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.662755 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.685050 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.689050 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 23:52:12.695636297 +0000 UTC Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.722136 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.742348 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.752037 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.752104 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.752126 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.752152 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.752177 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:45Z","lastTransitionTime":"2026-02-20T08:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.761769 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.778388 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6cfa24bc2ee3daae226cdfb1ec830d8a97c86eed37ca8ce2d0c10878627bcc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cc8878e457802d4970c4db52fc332cf96d26ab80c90cbe1c3160b8c02b385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.793263 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xd86g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xd86g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.826916 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:26Z\\\",\\\"message\\\":\\\"rmers/externalversions/factory.go:141\\\\nI0220 08:06:26.064417 6421 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:06:26.065536 6421 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0220 08:06:26.065606 6421 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 08:06:26.065641 6421 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 08:06:26.065650 6421 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 08:06:26.065682 6421 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0220 08:06:26.065678 6421 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0220 08:06:26.065713 6421 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 08:06:26.065690 6421 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0220 08:06:26.065709 6421 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 08:06:26.065775 6421 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 08:06:26.065788 6421 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0220 08:06:26.065794 6421 factory.go:656] Stopping watch factory\\\\nI0220 08:06:26.065810 6421 handler.go:208] Removed *v1.Pod ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:44Z\\\",\\\"message\\\":\\\"ate admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z]\\\\nI0220 08:06:44.892233 6647 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/community-operators]} name:Service_openshift-marketplace/community-operators_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.189:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d389393c-7ba9-422c-b3f5-06e391d537d2}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.850420 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.855442 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.855542 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.855569 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.855605 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.855628 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:45Z","lastTransitionTime":"2026-02-20T08:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.876491 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.897314 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.916821 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.937667 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.953887 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ae14aa1-2ba5-4630-b0b9-97cef2fece9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://001bf218b5793a1101d9960b6d81173d67c5f6259c21d780623dad9808adfd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5f1450fc3f715e0082ad316af3f03ca0d2d8f32b5efa9850e0694864f79153e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a52eab83c2f040b6bd090f63f55a7da942f1b9a219ae29f14fc2b9930b07da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.961345 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.961382 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.961394 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.961414 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.961428 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:45Z","lastTransitionTime":"2026-02-20T08:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.975731 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.991512 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.065517 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.065578 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.065590 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.065613 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.065627 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:46Z","lastTransitionTime":"2026-02-20T08:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.170209 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.170283 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.170302 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.170332 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.170351 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:46Z","lastTransitionTime":"2026-02-20T08:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.194212 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovnkube-controller/2.log" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.200771 4948 scope.go:117] "RemoveContainer" containerID="bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf" Feb 20 08:06:46 crc kubenswrapper[4948]: E0220 08:06:46.201108 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.218474 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6cfa24bc2ee3daae226cdfb1ec830d8a97c86eed37ca8ce2d0c10878627bcc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cc8878e457802d4970c4db52fc332cf96d26ab80c90cbe1c3160b8c02b385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.236593 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xd86g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xd86g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.274039 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.274091 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.274104 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.274129 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.274145 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:46Z","lastTransitionTime":"2026-02-20T08:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.278554 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.301196 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.318736 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.344385 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.362795 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.377035 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.377099 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.377120 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.377148 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.377170 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:46Z","lastTransitionTime":"2026-02-20T08:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.396861 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:44Z\\\",\\\"message\\\":\\\"ate admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z]\\\\nI0220 08:06:44.892233 6647 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/community-operators]} name:Service_openshift-marketplace/community-operators_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.189:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d389393c-7ba9-422c-b3f5-06e391d537d2}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.419941 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.436070 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.452946 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.476276 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.480387 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.480433 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.480449 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.480473 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.480489 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:46Z","lastTransitionTime":"2026-02-20T08:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.496561 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ae14aa1-2ba5-4630-b0b9-97cef2fece9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://001bf218b5793a1101d9960b6d81173d67c5f6259c21d780623dad9808adfd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5f1450fc3f715e0082ad316af3f03ca0d2d8f32b5efa9850e0694864f79153e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a52eab83c2f040b6bd090f63f55a7da942f1b9a219ae29f14fc2b9930b07da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.517474 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.536131 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.554702 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.575267 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.583471 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.583595 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.583611 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.583633 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.583646 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:46Z","lastTransitionTime":"2026-02-20T08:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.593394 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.686808 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.686891 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.686911 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.686950 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.687011 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:46Z","lastTransitionTime":"2026-02-20T08:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.689967 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 18:24:17.177828311 +0000 UTC Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.721434 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.721494 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.721454 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.721434 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:46 crc kubenswrapper[4948]: E0220 08:06:46.721602 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:06:46 crc kubenswrapper[4948]: E0220 08:06:46.721828 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:46 crc kubenswrapper[4948]: E0220 08:06:46.721966 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:46 crc kubenswrapper[4948]: E0220 08:06:46.722111 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.790531 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.790602 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.790621 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.790646 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.790661 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:46Z","lastTransitionTime":"2026-02-20T08:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.893853 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.893924 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.893944 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.894003 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.894022 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:46Z","lastTransitionTime":"2026-02-20T08:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.996795 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.996851 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.996867 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.996893 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.996910 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:46Z","lastTransitionTime":"2026-02-20T08:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.100662 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.100727 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.100751 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.100794 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.100815 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:47Z","lastTransitionTime":"2026-02-20T08:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.203821 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.203891 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.203915 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.203945 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.204016 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:47Z","lastTransitionTime":"2026-02-20T08:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.307305 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.307431 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.307452 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.307476 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.307495 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:47Z","lastTransitionTime":"2026-02-20T08:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.410736 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.410804 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.410827 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.410858 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.410881 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:47Z","lastTransitionTime":"2026-02-20T08:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.514152 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.514237 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.514258 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.514285 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.514303 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:47Z","lastTransitionTime":"2026-02-20T08:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.617674 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.617745 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.617763 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.617790 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.617809 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:47Z","lastTransitionTime":"2026-02-20T08:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.691104 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 06:26:23.148617268 +0000 UTC Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.720831 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.720882 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.720901 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.720929 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.720949 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:47Z","lastTransitionTime":"2026-02-20T08:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.829579 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.829662 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.829680 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.829711 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.829729 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:47Z","lastTransitionTime":"2026-02-20T08:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.933655 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.933726 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.933746 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.933777 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.933797 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:47Z","lastTransitionTime":"2026-02-20T08:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.037291 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.037392 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.037412 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.037445 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.037465 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:48Z","lastTransitionTime":"2026-02-20T08:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.141304 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.141365 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.141383 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.141409 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.141431 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:48Z","lastTransitionTime":"2026-02-20T08:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.245191 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.245236 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.245250 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.245269 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.245284 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:48Z","lastTransitionTime":"2026-02-20T08:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.349334 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.349436 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.349454 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.349535 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.349557 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:48Z","lastTransitionTime":"2026-02-20T08:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.452864 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.452955 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.453028 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.453064 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.453091 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:48Z","lastTransitionTime":"2026-02-20T08:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.556947 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.557137 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.557170 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.557203 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.557227 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:48Z","lastTransitionTime":"2026-02-20T08:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.660780 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.660844 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.660862 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.660887 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.660906 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:48Z","lastTransitionTime":"2026-02-20T08:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.691887 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 01:30:44.263786704 +0000 UTC Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.722655 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.722704 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.722665 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.722656 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:48 crc kubenswrapper[4948]: E0220 08:06:48.722812 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:48 crc kubenswrapper[4948]: E0220 08:06:48.723024 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:06:48 crc kubenswrapper[4948]: E0220 08:06:48.723158 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:48 crc kubenswrapper[4948]: E0220 08:06:48.723304 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.764649 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.764741 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.764767 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.764800 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.764825 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:48Z","lastTransitionTime":"2026-02-20T08:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.868296 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.868382 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.868401 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.868459 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.868480 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:48Z","lastTransitionTime":"2026-02-20T08:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.972283 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.972727 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.972827 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.972939 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.973062 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:48Z","lastTransitionTime":"2026-02-20T08:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.076117 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.076158 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.076170 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.076187 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.076202 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:49Z","lastTransitionTime":"2026-02-20T08:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.180332 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.180411 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.180433 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.180467 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.180489 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:49Z","lastTransitionTime":"2026-02-20T08:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.283496 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.283546 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.283566 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.283588 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.283610 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:49Z","lastTransitionTime":"2026-02-20T08:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.386282 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.386343 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.386367 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.386394 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.386417 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:49Z","lastTransitionTime":"2026-02-20T08:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.399619 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.399777 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.399805 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.399832 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.399853 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:49Z","lastTransitionTime":"2026-02-20T08:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:49 crc kubenswrapper[4948]: E0220 08:06:49.421716 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:49Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.428182 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.428245 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.428268 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.428295 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.428316 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:49Z","lastTransitionTime":"2026-02-20T08:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:49 crc kubenswrapper[4948]: E0220 08:06:49.451192 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:49Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.457599 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.457654 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.457671 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.457699 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.457718 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:49Z","lastTransitionTime":"2026-02-20T08:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:49 crc kubenswrapper[4948]: E0220 08:06:49.479788 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:49Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.485126 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.485170 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.485197 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.485229 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.485255 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:49Z","lastTransitionTime":"2026-02-20T08:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:49 crc kubenswrapper[4948]: E0220 08:06:49.504939 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:49Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.510014 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.510055 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.510073 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.510096 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.510114 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:49Z","lastTransitionTime":"2026-02-20T08:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:49 crc kubenswrapper[4948]: E0220 08:06:49.528740 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:49Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:49 crc kubenswrapper[4948]: E0220 08:06:49.529127 4948 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.530916 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.531025 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.531044 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.531064 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.531081 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:49Z","lastTransitionTime":"2026-02-20T08:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.634355 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.634426 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.634446 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.634474 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.634495 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:49Z","lastTransitionTime":"2026-02-20T08:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.693065 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 13:48:21.287241825 +0000 UTC Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.737319 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.737404 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.737431 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.737462 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.737487 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:49Z","lastTransitionTime":"2026-02-20T08:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.840577 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.840642 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.840661 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.840689 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.840711 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:49Z","lastTransitionTime":"2026-02-20T08:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.944186 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.944254 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.944278 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.944308 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.944331 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:49Z","lastTransitionTime":"2026-02-20T08:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.047520 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.047580 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.047603 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.047629 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.047649 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:50Z","lastTransitionTime":"2026-02-20T08:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.150832 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.150900 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.150923 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.150949 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.151007 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:50Z","lastTransitionTime":"2026-02-20T08:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.253319 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.253392 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.253411 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.253440 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.253460 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:50Z","lastTransitionTime":"2026-02-20T08:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.356637 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.356702 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.356731 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.356763 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.356786 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:50Z","lastTransitionTime":"2026-02-20T08:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.460260 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.460326 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.460339 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.460358 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.460370 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:50Z","lastTransitionTime":"2026-02-20T08:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.563494 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.563552 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.563570 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.563593 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.563611 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:50Z","lastTransitionTime":"2026-02-20T08:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.666252 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.666457 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.666483 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.666506 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.666523 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:50Z","lastTransitionTime":"2026-02-20T08:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.694221 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 02:06:40.87345129 +0000 UTC Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.721837 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.721906 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.721871 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.721922 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:50 crc kubenswrapper[4948]: E0220 08:06:50.722088 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:50 crc kubenswrapper[4948]: E0220 08:06:50.722409 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:06:50 crc kubenswrapper[4948]: E0220 08:06:50.722524 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:50 crc kubenswrapper[4948]: E0220 08:06:50.722658 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.769520 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.769611 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.769638 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.769672 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.769699 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:50Z","lastTransitionTime":"2026-02-20T08:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.873562 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.873619 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.873635 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.873659 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.873675 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:50Z","lastTransitionTime":"2026-02-20T08:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.977040 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.977118 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.977136 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.977167 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.977191 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:50Z","lastTransitionTime":"2026-02-20T08:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.080480 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.080816 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.081072 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.081226 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.081377 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:51Z","lastTransitionTime":"2026-02-20T08:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.184177 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.184258 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.184276 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.184301 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.184319 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:51Z","lastTransitionTime":"2026-02-20T08:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.287583 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.287678 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.287708 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.287740 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.287759 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:51Z","lastTransitionTime":"2026-02-20T08:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.391591 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.391654 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.391675 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.391704 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.391726 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:51Z","lastTransitionTime":"2026-02-20T08:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.495622 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.495695 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.495715 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.495747 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.495764 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:51Z","lastTransitionTime":"2026-02-20T08:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.599451 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.599673 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.599697 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.599723 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.599741 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:51Z","lastTransitionTime":"2026-02-20T08:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.695130 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 01:55:40.955891129 +0000 UTC Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.703099 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.703319 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.703476 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.703630 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.703771 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:51Z","lastTransitionTime":"2026-02-20T08:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.760765 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:51Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.782739 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:51Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.799904 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:51Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.806729 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.806803 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.806822 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.806853 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.806872 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:51Z","lastTransitionTime":"2026-02-20T08:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.818374 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6cfa24bc2ee3daae226cdfb1ec830d8a97c86eed37ca8ce2d0c10878627bcc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cc8878e457802d4970c4db52fc332cf96d26ab80c90cbe1c3160b8c02b385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:51Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.838055 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xd86g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xd86g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:51Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.862754 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:51Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.882920 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:51Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.908803 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:51Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.911836 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.911902 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.911922 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.911949 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.911968 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:51Z","lastTransitionTime":"2026-02-20T08:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.929101 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:51Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.966762 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:44Z\\\",\\\"message\\\":\\\"ate admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z]\\\\nI0220 08:06:44.892233 6647 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/community-operators]} name:Service_openshift-marketplace/community-operators_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.189:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d389393c-7ba9-422c-b3f5-06e391d537d2}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:51Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.984200 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:51Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.003135 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ae14aa1-2ba5-4630-b0b9-97cef2fece9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://001bf218b5793a1101d9960b6d81173d67c5f6259c21d780623dad9808adfd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5f1450fc3f715e0082ad316af3f03ca0d2d8f32b5efa9850e0694864f79153e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a52eab83c2f040b6bd090f63f55a7da942f1b9a219ae29f14fc2b9930b07da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:52Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.016654 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.016721 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.016752 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.016786 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.016805 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:52Z","lastTransitionTime":"2026-02-20T08:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.027475 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:52Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.045314 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:52Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.063306 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:52Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.081737 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:52Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.098555 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:52Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.115289 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:52Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.119588 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.119662 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.119683 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.119711 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.119730 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:52Z","lastTransitionTime":"2026-02-20T08:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.223348 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.223443 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.223464 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.223498 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.223519 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:52Z","lastTransitionTime":"2026-02-20T08:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.327013 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.327068 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.327086 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.327112 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.327175 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:52Z","lastTransitionTime":"2026-02-20T08:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.430078 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.430144 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.430162 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.430187 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.430210 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:52Z","lastTransitionTime":"2026-02-20T08:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.534116 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.534180 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.534245 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.534273 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.534295 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:52Z","lastTransitionTime":"2026-02-20T08:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.637744 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.637792 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.637810 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.637837 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.637855 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:52Z","lastTransitionTime":"2026-02-20T08:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.696249 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 08:47:54.079079862 +0000 UTC Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.722124 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.722167 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.722299 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.722395 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:52 crc kubenswrapper[4948]: E0220 08:06:52.722389 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:52 crc kubenswrapper[4948]: E0220 08:06:52.722584 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:52 crc kubenswrapper[4948]: E0220 08:06:52.722790 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:06:52 crc kubenswrapper[4948]: E0220 08:06:52.723029 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.741295 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.741357 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.741375 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.741401 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.741420 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:52Z","lastTransitionTime":"2026-02-20T08:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.844849 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.844925 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.844943 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.845006 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.845026 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:52Z","lastTransitionTime":"2026-02-20T08:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.948502 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.949026 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.949219 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.949383 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.949516 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:52Z","lastTransitionTime":"2026-02-20T08:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.053556 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.053602 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.053617 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.053638 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.053653 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:53Z","lastTransitionTime":"2026-02-20T08:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.157204 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.157276 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.157295 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.157323 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.157342 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:53Z","lastTransitionTime":"2026-02-20T08:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.261047 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.261527 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.261734 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.261926 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.262159 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:53Z","lastTransitionTime":"2026-02-20T08:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.366388 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.366458 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.366478 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.366507 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.366526 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:53Z","lastTransitionTime":"2026-02-20T08:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.470016 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.470328 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.470508 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.470624 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.470713 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:53Z","lastTransitionTime":"2026-02-20T08:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.573911 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.574073 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.574107 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.574144 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.574171 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:53Z","lastTransitionTime":"2026-02-20T08:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.677901 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.678025 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.678046 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.678081 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.678102 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:53Z","lastTransitionTime":"2026-02-20T08:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.697475 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 18:22:32.438410222 +0000 UTC Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.782317 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.782404 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.782445 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.782480 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.782502 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:53Z","lastTransitionTime":"2026-02-20T08:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.885381 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.885437 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.885453 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.885475 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.885492 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:53Z","lastTransitionTime":"2026-02-20T08:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.988623 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.988682 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.988698 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.988723 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.988741 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:53Z","lastTransitionTime":"2026-02-20T08:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.092031 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.092090 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.092117 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.092139 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.092155 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:54Z","lastTransitionTime":"2026-02-20T08:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.194766 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.194855 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.194878 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.194906 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.194923 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:54Z","lastTransitionTime":"2026-02-20T08:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.298109 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.298274 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.298299 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.298326 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.298344 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:54Z","lastTransitionTime":"2026-02-20T08:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.402949 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.403041 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.403061 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.403088 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.403109 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:54Z","lastTransitionTime":"2026-02-20T08:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.506815 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.506868 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.506885 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.506913 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.506933 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:54Z","lastTransitionTime":"2026-02-20T08:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.610967 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.611068 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.611090 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.611117 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.611138 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:54Z","lastTransitionTime":"2026-02-20T08:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.697994 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 01:45:21.545583545 +0000 UTC Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.714927 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.715047 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.715074 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.715106 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.715130 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:54Z","lastTransitionTime":"2026-02-20T08:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.722202 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.722259 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.722290 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.722202 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:54 crc kubenswrapper[4948]: E0220 08:06:54.722402 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:06:54 crc kubenswrapper[4948]: E0220 08:06:54.722500 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:54 crc kubenswrapper[4948]: E0220 08:06:54.722625 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:54 crc kubenswrapper[4948]: E0220 08:06:54.722746 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.818708 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.818775 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.818798 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.818829 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.818851 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:54Z","lastTransitionTime":"2026-02-20T08:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.922459 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.922515 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.922532 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.922556 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.922577 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:54Z","lastTransitionTime":"2026-02-20T08:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.025585 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.025662 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.025680 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.025708 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.025727 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:55Z","lastTransitionTime":"2026-02-20T08:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.128928 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.129034 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.129057 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.129082 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.129104 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:55Z","lastTransitionTime":"2026-02-20T08:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.232527 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.232595 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.232615 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.232640 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.232659 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:55Z","lastTransitionTime":"2026-02-20T08:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.336493 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.336556 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.336577 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.336601 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.336621 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:55Z","lastTransitionTime":"2026-02-20T08:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.438906 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.438952 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.438963 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.438997 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.439010 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:55Z","lastTransitionTime":"2026-02-20T08:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.541652 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.541761 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.541785 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.541810 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.541828 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:55Z","lastTransitionTime":"2026-02-20T08:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.645383 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.645430 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.645446 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.645464 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.645476 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:55Z","lastTransitionTime":"2026-02-20T08:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.698316 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 00:11:58.672805079 +0000 UTC Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.748449 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.748522 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.748545 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.748576 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.748595 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:55Z","lastTransitionTime":"2026-02-20T08:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.850999 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.851051 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.851063 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.851082 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.851095 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:55Z","lastTransitionTime":"2026-02-20T08:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.954071 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.954131 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.954141 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.954159 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.954169 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:55Z","lastTransitionTime":"2026-02-20T08:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.056749 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.056820 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.056839 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.056869 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.056891 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:56Z","lastTransitionTime":"2026-02-20T08:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.161480 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.161542 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.161555 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.161578 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.161595 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:56Z","lastTransitionTime":"2026-02-20T08:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.265014 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.265065 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.265075 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.265407 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.265422 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:56Z","lastTransitionTime":"2026-02-20T08:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.369563 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.369665 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.369684 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.369709 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.369727 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:56Z","lastTransitionTime":"2026-02-20T08:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.473277 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.473690 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.473865 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.474091 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.474308 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:56Z","lastTransitionTime":"2026-02-20T08:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.577540 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.577634 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.577651 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.577674 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.577693 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:56Z","lastTransitionTime":"2026-02-20T08:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.680963 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.681031 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.681041 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.681057 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.681070 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:56Z","lastTransitionTime":"2026-02-20T08:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.699393 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 23:16:06.111703225 +0000 UTC Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.721953 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:56 crc kubenswrapper[4948]: E0220 08:06:56.722167 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.722256 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.722346 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:56 crc kubenswrapper[4948]: E0220 08:06:56.722373 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.722460 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:56 crc kubenswrapper[4948]: E0220 08:06:56.722558 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:56 crc kubenswrapper[4948]: E0220 08:06:56.722696 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.784127 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.784162 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.784170 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.784184 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.784194 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:56Z","lastTransitionTime":"2026-02-20T08:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.886653 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.886711 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.886729 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.886773 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.886792 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:56Z","lastTransitionTime":"2026-02-20T08:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.989170 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.989241 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.989264 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.989293 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.989326 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:56Z","lastTransitionTime":"2026-02-20T08:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.092919 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.093008 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.093018 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.093040 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.093051 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:57Z","lastTransitionTime":"2026-02-20T08:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.195928 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.196009 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.196026 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.196048 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.196065 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:57Z","lastTransitionTime":"2026-02-20T08:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.299103 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.299156 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.299166 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.299186 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.299197 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:57Z","lastTransitionTime":"2026-02-20T08:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.402711 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.402776 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.402820 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.402846 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.402863 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:57Z","lastTransitionTime":"2026-02-20T08:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.506423 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.506478 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.506497 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.506521 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.506539 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:57Z","lastTransitionTime":"2026-02-20T08:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.610299 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.610344 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.610360 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.610384 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.610400 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:57Z","lastTransitionTime":"2026-02-20T08:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.699615 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 12:05:56.474940802 +0000 UTC Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.713723 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.713817 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.713845 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.713880 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.713906 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:57Z","lastTransitionTime":"2026-02-20T08:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.817278 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.817334 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.817344 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.817359 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.817369 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:57Z","lastTransitionTime":"2026-02-20T08:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.920639 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.920707 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.920732 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.920757 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.920776 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:57Z","lastTransitionTime":"2026-02-20T08:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.023489 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.023554 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.023578 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.023607 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.023631 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:58Z","lastTransitionTime":"2026-02-20T08:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.126701 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.126763 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.126786 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.126812 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.126831 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:58Z","lastTransitionTime":"2026-02-20T08:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.230685 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.230744 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.230762 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.230787 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.230806 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:58Z","lastTransitionTime":"2026-02-20T08:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.360712 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.360749 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.360760 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.360779 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.360793 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:58Z","lastTransitionTime":"2026-02-20T08:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.464857 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.464920 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.464934 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.464952 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.464989 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:58Z","lastTransitionTime":"2026-02-20T08:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.568205 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.568246 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.568255 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.568274 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.568284 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:58Z","lastTransitionTime":"2026-02-20T08:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.671670 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.671717 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.671727 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.671745 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.671755 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:58Z","lastTransitionTime":"2026-02-20T08:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.699920 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 00:49:32.57491579 +0000 UTC Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.722338 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.722340 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.722368 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:58 crc kubenswrapper[4948]: E0220 08:06:58.722539 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.722553 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:58 crc kubenswrapper[4948]: E0220 08:06:58.722653 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:58 crc kubenswrapper[4948]: E0220 08:06:58.722734 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:58 crc kubenswrapper[4948]: E0220 08:06:58.722796 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.734676 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.773636 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.773672 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.773680 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.773693 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.773703 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:58Z","lastTransitionTime":"2026-02-20T08:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.886016 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.886100 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.886127 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.886159 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.886183 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:58Z","lastTransitionTime":"2026-02-20T08:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.988481 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.988560 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.988576 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.988596 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.988610 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:58Z","lastTransitionTime":"2026-02-20T08:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.091017 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.091093 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.091105 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.091126 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.091142 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:59Z","lastTransitionTime":"2026-02-20T08:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.193220 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.193251 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.193263 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.193280 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.193290 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:59Z","lastTransitionTime":"2026-02-20T08:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.295094 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.295135 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.295147 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.295162 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.295186 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:59Z","lastTransitionTime":"2026-02-20T08:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.397927 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.398030 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.398046 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.398065 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.398080 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:59Z","lastTransitionTime":"2026-02-20T08:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.501057 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.501099 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.501108 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.501125 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.501135 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:59Z","lastTransitionTime":"2026-02-20T08:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.592211 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.592255 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.592264 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.592280 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.592291 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:59Z","lastTransitionTime":"2026-02-20T08:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:59 crc kubenswrapper[4948]: E0220 08:06:59.611307 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:59Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.615318 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.615374 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.615389 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.615411 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.615428 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:59Z","lastTransitionTime":"2026-02-20T08:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:59 crc kubenswrapper[4948]: E0220 08:06:59.626858 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:59Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.631680 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.631730 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.631740 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.631755 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.631765 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:59Z","lastTransitionTime":"2026-02-20T08:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:59 crc kubenswrapper[4948]: E0220 08:06:59.644775 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:59Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.650119 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.650235 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.650254 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.650280 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.650301 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:59Z","lastTransitionTime":"2026-02-20T08:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:59 crc kubenswrapper[4948]: E0220 08:06:59.669634 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:59Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.674678 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.674758 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.674772 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.674794 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.674874 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:59Z","lastTransitionTime":"2026-02-20T08:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:59 crc kubenswrapper[4948]: E0220 08:06:59.688111 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:59Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:59 crc kubenswrapper[4948]: E0220 08:06:59.688239 4948 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.689856 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.689898 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.689912 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.689931 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.689945 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:59Z","lastTransitionTime":"2026-02-20T08:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.700288 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 13:19:15.391288625 +0000 UTC Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.792590 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.792663 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.792679 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.792700 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.792719 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:59Z","lastTransitionTime":"2026-02-20T08:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.894738 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.894777 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.894808 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.894822 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.894832 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:59Z","lastTransitionTime":"2026-02-20T08:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.997226 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.997269 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.997283 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.997297 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.997311 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:59Z","lastTransitionTime":"2026-02-20T08:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.100253 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.100301 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.100314 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.100337 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.100352 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:00Z","lastTransitionTime":"2026-02-20T08:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.203225 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.203294 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.203318 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.203349 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.203372 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:00Z","lastTransitionTime":"2026-02-20T08:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.305935 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.305992 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.306005 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.306026 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.306040 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:00Z","lastTransitionTime":"2026-02-20T08:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.408894 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.408927 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.408939 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.408954 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.408966 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:00Z","lastTransitionTime":"2026-02-20T08:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.511735 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.511766 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.511777 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.511790 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.511799 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:00Z","lastTransitionTime":"2026-02-20T08:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.614310 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.614370 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.614391 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.614422 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.614442 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:00Z","lastTransitionTime":"2026-02-20T08:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.681295 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs\") pod \"network-metrics-daemon-xd86g\" (UID: \"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\") " pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:00 crc kubenswrapper[4948]: E0220 08:07:00.681498 4948 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 08:07:00 crc kubenswrapper[4948]: E0220 08:07:00.681588 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs podName:62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd nodeName:}" failed. No retries permitted until 2026-02-20 08:07:32.681563292 +0000 UTC m=+101.656058112 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs") pod "network-metrics-daemon-xd86g" (UID: "62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.700597 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 23:56:33.573767755 +0000 UTC Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.716769 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.716807 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.716818 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.716835 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.716848 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:00Z","lastTransitionTime":"2026-02-20T08:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.722386 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.722449 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:00 crc kubenswrapper[4948]: E0220 08:07:00.722529 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.722629 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:00 crc kubenswrapper[4948]: E0220 08:07:00.722776 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.722850 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:00 crc kubenswrapper[4948]: E0220 08:07:00.722949 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:00 crc kubenswrapper[4948]: E0220 08:07:00.723477 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.723935 4948 scope.go:117] "RemoveContainer" containerID="bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf" Feb 20 08:07:00 crc kubenswrapper[4948]: E0220 08:07:00.724195 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.821599 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.821665 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.821686 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.821714 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.821752 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:00Z","lastTransitionTime":"2026-02-20T08:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.924158 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.924208 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.924222 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.924244 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.924257 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:00Z","lastTransitionTime":"2026-02-20T08:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.027735 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.027782 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.027796 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.027817 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.027828 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:01Z","lastTransitionTime":"2026-02-20T08:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.131245 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.131292 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.131301 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.131317 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.131327 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:01Z","lastTransitionTime":"2026-02-20T08:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.233597 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.233631 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.233642 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.233656 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.233666 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:01Z","lastTransitionTime":"2026-02-20T08:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.335944 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.336005 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.336014 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.336032 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.336044 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:01Z","lastTransitionTime":"2026-02-20T08:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.438488 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.438522 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.438535 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.438556 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.438568 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:01Z","lastTransitionTime":"2026-02-20T08:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.540839 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.540878 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.540892 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.540911 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.540926 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:01Z","lastTransitionTime":"2026-02-20T08:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.643623 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.643656 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.643666 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.643681 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.643691 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:01Z","lastTransitionTime":"2026-02-20T08:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.701410 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 17:25:14.277958375 +0000 UTC Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.741926 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:01Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.745501 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.745521 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.745529 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.745539 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.745547 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:01Z","lastTransitionTime":"2026-02-20T08:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.761702 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ae14aa1-2ba5-4630-b0b9-97cef2fece9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://001bf218b5793a1101d9960b6d81173d67c5f6259c21d780623dad9808adfd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5f1450fc3f715e0082ad316af3f03ca0d2d8f32b5efa9850e0694864f79153e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a52eab83c2f040b6bd090f63f55a7da942f1b9a219ae29f14fc2b9930b07da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:01Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.783660 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:01Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.801026 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:01Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.820763 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:01Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.839963 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:01Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.848340 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.848396 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.848409 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.848433 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.848450 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:01Z","lastTransitionTime":"2026-02-20T08:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.859063 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:01Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.881729 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:01Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.915128 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:01Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.934591 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:01Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.949645 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:01Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.950758 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.950809 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.950820 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.950837 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.950847 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:01Z","lastTransitionTime":"2026-02-20T08:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.963839 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6cfa24bc2ee3daae226cdfb1ec830d8a97c86eed37ca8ce2d0c10878627bcc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cc8878e457802d4970c4db52fc332cf96d26ab80c90cbe1c3160b8c02b385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:01Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.974662 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xd86g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xd86g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:01Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.983906 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ad7145-074b-49df-a0e9-cfec6e3974ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67b60ad2efa00639a153f39923bfdba3b40b6b3f5126207b4f0e05abdf9f43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:01Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.998507 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:01Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.009269 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.025953 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.040135 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.052938 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.053001 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.053015 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.053031 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.053045 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:02Z","lastTransitionTime":"2026-02-20T08:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.067953 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:44Z\\\",\\\"message\\\":\\\"ate admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z]\\\\nI0220 08:06:44.892233 6647 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/community-operators]} name:Service_openshift-marketplace/community-operators_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.189:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d389393c-7ba9-422c-b3f5-06e391d537d2}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.154944 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.155018 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.155035 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.155056 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.155074 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:02Z","lastTransitionTime":"2026-02-20T08:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.258986 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.259358 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.259367 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.259385 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.259397 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:02Z","lastTransitionTime":"2026-02-20T08:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.262493 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x7s56_b96124e4-0a74-4578-9142-fd728eb9f99e/kube-multus/0.log" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.262539 4948 generic.go:334] "Generic (PLEG): container finished" podID="b96124e4-0a74-4578-9142-fd728eb9f99e" containerID="34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e" exitCode=1 Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.262583 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x7s56" event={"ID":"b96124e4-0a74-4578-9142-fd728eb9f99e","Type":"ContainerDied","Data":"34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e"} Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.263583 4948 scope.go:117] "RemoveContainer" containerID="34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.287623 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.305246 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ae14aa1-2ba5-4630-b0b9-97cef2fece9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://001bf218b5793a1101d9960b6d81173d67c5f6259c21d780623dad9808adfd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5f1450fc3f715e0082ad316af3f03ca0d2d8f32b5efa9850e0694864f79153e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a52eab83c2f040b6bd090f63f55a7da942f1b9a219ae29f14fc2b9930b07da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.319213 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.331768 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.345649 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.359243 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.368243 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.368309 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.368328 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.368357 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.368378 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:02Z","lastTransitionTime":"2026-02-20T08:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.372728 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.392939 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:07:02Z\\\",\\\"message\\\":\\\"2026-02-20T08:06:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b42b7938-8525-4811-9d9f-355344db9d45\\\\n2026-02-20T08:06:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b42b7938-8525-4811-9d9f-355344db9d45 to /host/opt/cni/bin/\\\\n2026-02-20T08:06:17Z [verbose] multus-daemon started\\\\n2026-02-20T08:06:17Z [verbose] Readiness Indicator file check\\\\n2026-02-20T08:07:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.413898 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.427925 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.441785 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.457595 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6cfa24bc2ee3daae226cdfb1ec830d8a97c86eed37ca8ce2d0c10878627bcc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cc8878e457802d4970c4db52fc332cf96d26ab80c90cbe1c3160b8c02b385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.469396 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xd86g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xd86g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.471298 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.471337 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.471345 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.471363 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.471377 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:02Z","lastTransitionTime":"2026-02-20T08:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.480287 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ad7145-074b-49df-a0e9-cfec6e3974ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67b60ad2efa00639a153f39923bfdba3b40b6b3f5126207b4f0e05abdf9f43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.494270 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.504869 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.517265 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.533515 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.558846 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:44Z\\\",\\\"message\\\":\\\"ate admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z]\\\\nI0220 08:06:44.892233 6647 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/community-operators]} name:Service_openshift-marketplace/community-operators_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.189:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d389393c-7ba9-422c-b3f5-06e391d537d2}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.573850 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.573905 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.573924 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.573950 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.573996 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:02Z","lastTransitionTime":"2026-02-20T08:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.677653 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.677712 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.677729 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.677749 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.677761 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:02Z","lastTransitionTime":"2026-02-20T08:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.702288 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 23:09:13.973968472 +0000 UTC Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.721890 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.721930 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.721904 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.721894 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:02 crc kubenswrapper[4948]: E0220 08:07:02.722076 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:02 crc kubenswrapper[4948]: E0220 08:07:02.722220 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:02 crc kubenswrapper[4948]: E0220 08:07:02.722390 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:02 crc kubenswrapper[4948]: E0220 08:07:02.722523 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.780360 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.780424 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.780443 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.780471 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.780490 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:02Z","lastTransitionTime":"2026-02-20T08:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.883899 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.883949 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.883958 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.883988 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.884000 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:02Z","lastTransitionTime":"2026-02-20T08:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.986915 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.987265 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.987328 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.987393 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.987460 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:02Z","lastTransitionTime":"2026-02-20T08:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.090710 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.090805 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.090831 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.090865 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.090893 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:03Z","lastTransitionTime":"2026-02-20T08:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.194921 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.195106 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.195225 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.196131 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.196196 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:03Z","lastTransitionTime":"2026-02-20T08:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.269939 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x7s56_b96124e4-0a74-4578-9142-fd728eb9f99e/kube-multus/0.log" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.270042 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x7s56" event={"ID":"b96124e4-0a74-4578-9142-fd728eb9f99e","Type":"ContainerStarted","Data":"2f4093ec9d71181c58ad80a8c4ee3692148590df14aadaec96c0f54a843d1f8c"} Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.288715 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ae14aa1-2ba5-4630-b0b9-97cef2fece9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://001bf218b5793a1101d9960b6d81173d67c5f6259c21d780623dad9808adfd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5f1450fc3f715e0082ad316af3f03ca0d2d8f32b5efa9850e0694864f79153e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a52eab83c2f040b6bd090f63f55a7da942f1b9a219ae29f14fc2b9930b07da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.298754 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.298807 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.298831 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.298858 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.298878 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:03Z","lastTransitionTime":"2026-02-20T08:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.315243 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.331629 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.344559 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.358511 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.374803 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.394186 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f4093ec9d71181c58ad80a8c4ee3692148590df14aadaec96c0f54a843d1f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:07:02Z\\\",\\\"message\\\":\\\"2026-02-20T08:06:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b42b7938-8525-4811-9d9f-355344db9d45\\\\n2026-02-20T08:06:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b42b7938-8525-4811-9d9f-355344db9d45 to /host/opt/cni/bin/\\\\n2026-02-20T08:06:17Z [verbose] multus-daemon started\\\\n2026-02-20T08:06:17Z [verbose] Readiness Indicator file check\\\\n2026-02-20T08:07:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.401156 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.401189 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.401220 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.401237 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.401250 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:03Z","lastTransitionTime":"2026-02-20T08:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.413661 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.430006 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.441241 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.454517 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6cfa24bc2ee3daae226cdfb1ec830d8a97c86eed37ca8ce2d0c10878627bcc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cc8878e457802d4970c4db52fc332cf96d26ab80c90cbe1c3160b8c02b385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.464803 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xd86g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xd86g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.487129 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.501749 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.503371 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.503424 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.503444 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.503467 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.503484 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:03Z","lastTransitionTime":"2026-02-20T08:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.515344 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.533474 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.545487 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.567827 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:44Z\\\",\\\"message\\\":\\\"ate admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z]\\\\nI0220 08:06:44.892233 6647 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/community-operators]} name:Service_openshift-marketplace/community-operators_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.189:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d389393c-7ba9-422c-b3f5-06e391d537d2}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.581449 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ad7145-074b-49df-a0e9-cfec6e3974ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67b60ad2efa00639a153f39923bfdba3b40b6b3f5126207b4f0e05abdf9f43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.606083 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.606174 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.606192 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.606219 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.606237 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:03Z","lastTransitionTime":"2026-02-20T08:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.703083 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 18:34:25.93003713 +0000 UTC Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.709036 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.709074 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.709084 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.709101 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.709114 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:03Z","lastTransitionTime":"2026-02-20T08:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.811301 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.811365 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.811379 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.811398 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.811411 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:03Z","lastTransitionTime":"2026-02-20T08:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.915017 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.915067 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.915076 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.915091 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.915104 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:03Z","lastTransitionTime":"2026-02-20T08:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.018568 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.018637 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.018657 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.018680 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.018695 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:04Z","lastTransitionTime":"2026-02-20T08:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.121888 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.121947 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.121962 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.122000 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.122012 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:04Z","lastTransitionTime":"2026-02-20T08:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.224878 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.224923 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.224936 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.224955 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.224988 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:04Z","lastTransitionTime":"2026-02-20T08:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.326881 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.326923 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.326933 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.326950 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.326963 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:04Z","lastTransitionTime":"2026-02-20T08:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.429949 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.430584 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.430687 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.430845 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.430957 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:04Z","lastTransitionTime":"2026-02-20T08:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.533934 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.534017 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.534036 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.534061 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.534080 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:04Z","lastTransitionTime":"2026-02-20T08:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.994133 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.994171 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:04 crc kubenswrapper[4948]: E0220 08:07:04.994271 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.994133 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:04 crc kubenswrapper[4948]: E0220 08:07:04.994359 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.994425 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:04 crc kubenswrapper[4948]: E0220 08:07:04.994539 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:04 crc kubenswrapper[4948]: E0220 08:07:04.994599 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.994639 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 13:09:44.211385592 +0000 UTC Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.995343 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.995362 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.995372 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.995386 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.995397 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:04Z","lastTransitionTime":"2026-02-20T08:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.097915 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.097955 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.097997 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.098018 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.098035 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:05Z","lastTransitionTime":"2026-02-20T08:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.200292 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.200328 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.200344 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.200365 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.200380 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:05Z","lastTransitionTime":"2026-02-20T08:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.302808 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.302875 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.302885 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.302909 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.302921 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:05Z","lastTransitionTime":"2026-02-20T08:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.405793 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.405845 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.405864 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.405893 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.405913 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:05Z","lastTransitionTime":"2026-02-20T08:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.508047 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.508086 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.508096 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.508113 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.508128 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:05Z","lastTransitionTime":"2026-02-20T08:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.611256 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.611301 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.611314 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.611334 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.611346 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:05Z","lastTransitionTime":"2026-02-20T08:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.713609 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.713728 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.713756 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.713790 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.713818 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:05Z","lastTransitionTime":"2026-02-20T08:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.817063 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.817142 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.817168 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.817201 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.817231 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:05Z","lastTransitionTime":"2026-02-20T08:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.921366 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.921426 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.921445 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.921470 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.921487 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:05Z","lastTransitionTime":"2026-02-20T08:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.995464 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 00:06:08.18944778 +0000 UTC Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.024876 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.024936 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.024950 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.024988 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.025004 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:06Z","lastTransitionTime":"2026-02-20T08:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.127700 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.127755 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.127777 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.127808 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.127835 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:06Z","lastTransitionTime":"2026-02-20T08:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.231551 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.231614 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.231633 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.231657 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.231677 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:06Z","lastTransitionTime":"2026-02-20T08:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.335063 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.335129 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.335146 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.335169 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.335187 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:06Z","lastTransitionTime":"2026-02-20T08:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.438474 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.438544 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.438562 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.438586 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.438603 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:06Z","lastTransitionTime":"2026-02-20T08:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.541588 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.541651 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.541673 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.541703 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.541725 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:06Z","lastTransitionTime":"2026-02-20T08:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.645363 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.645423 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.645445 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.645477 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.645498 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:06Z","lastTransitionTime":"2026-02-20T08:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.721851 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:06 crc kubenswrapper[4948]: E0220 08:07:06.722139 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.722556 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:06 crc kubenswrapper[4948]: E0220 08:07:06.722710 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.723058 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:06 crc kubenswrapper[4948]: E0220 08:07:06.723217 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.725069 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:06 crc kubenswrapper[4948]: E0220 08:07:06.725247 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.749014 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.749054 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.749070 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.749096 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.749110 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:06Z","lastTransitionTime":"2026-02-20T08:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.852576 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.852613 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.852625 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.852641 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.852652 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:06Z","lastTransitionTime":"2026-02-20T08:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.955507 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.955577 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.955594 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.955612 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.955624 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:06Z","lastTransitionTime":"2026-02-20T08:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.996245 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 13:57:31.868688962 +0000 UTC Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.058595 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.058669 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.058707 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.058738 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.058761 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:07Z","lastTransitionTime":"2026-02-20T08:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.161639 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.161754 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.161793 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.161828 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.161857 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:07Z","lastTransitionTime":"2026-02-20T08:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.264821 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.264896 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.264915 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.264942 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.264961 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:07Z","lastTransitionTime":"2026-02-20T08:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.368618 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.368690 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.368708 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.368733 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.368752 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:07Z","lastTransitionTime":"2026-02-20T08:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.476926 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.477044 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.477070 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.477100 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.477122 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:07Z","lastTransitionTime":"2026-02-20T08:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.580685 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.580757 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.580774 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.580801 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.580820 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:07Z","lastTransitionTime":"2026-02-20T08:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.684552 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.684619 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.684642 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.684671 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.684694 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:07Z","lastTransitionTime":"2026-02-20T08:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.788072 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.788159 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.788174 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.788202 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.788221 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:07Z","lastTransitionTime":"2026-02-20T08:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.891106 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.891180 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.891199 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.891231 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.891250 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:07Z","lastTransitionTime":"2026-02-20T08:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.995113 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.995193 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.995217 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.995252 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.995277 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:07Z","lastTransitionTime":"2026-02-20T08:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.996471 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 06:00:40.051889535 +0000 UTC Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.097914 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.098017 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.098043 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.098074 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.098098 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:08Z","lastTransitionTime":"2026-02-20T08:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.201556 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.201601 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.201615 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.201635 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.201651 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:08Z","lastTransitionTime":"2026-02-20T08:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.303902 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.303957 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.304018 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.304043 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.304059 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:08Z","lastTransitionTime":"2026-02-20T08:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.407487 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.407552 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.407570 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.407597 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.407615 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:08Z","lastTransitionTime":"2026-02-20T08:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.511068 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.511502 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.511706 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.511898 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.512096 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:08Z","lastTransitionTime":"2026-02-20T08:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.615121 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.616069 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.616494 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.616925 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.617193 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:08Z","lastTransitionTime":"2026-02-20T08:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.720365 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.720427 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.720445 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.720473 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.720495 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:08Z","lastTransitionTime":"2026-02-20T08:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.721676 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.721737 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.721742 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:08 crc kubenswrapper[4948]: E0220 08:07:08.721830 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:08 crc kubenswrapper[4948]: E0220 08:07:08.722126 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:08 crc kubenswrapper[4948]: E0220 08:07:08.722150 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.721692 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:08 crc kubenswrapper[4948]: E0220 08:07:08.722676 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.823749 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.823816 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.823852 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.823891 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.823916 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:08Z","lastTransitionTime":"2026-02-20T08:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.927326 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.927375 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.927392 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.927414 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.927432 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:08Z","lastTransitionTime":"2026-02-20T08:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.997557 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 00:21:41.673377034 +0000 UTC Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.030808 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.030884 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.030912 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.030944 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.030968 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:09Z","lastTransitionTime":"2026-02-20T08:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.133490 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.133549 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.133566 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.133591 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.133609 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:09Z","lastTransitionTime":"2026-02-20T08:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.236635 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.236702 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.236721 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.236748 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.236767 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:09Z","lastTransitionTime":"2026-02-20T08:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.338832 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.338888 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.338904 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.338927 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.338944 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:09Z","lastTransitionTime":"2026-02-20T08:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.441960 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.442062 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.442087 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.442119 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.442139 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:09Z","lastTransitionTime":"2026-02-20T08:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.545184 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.545239 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.545251 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.545275 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.545296 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:09Z","lastTransitionTime":"2026-02-20T08:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.648390 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.648427 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.648436 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.648450 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.648461 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:09Z","lastTransitionTime":"2026-02-20T08:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.744456 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.744494 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.744502 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.744514 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.744524 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:09Z","lastTransitionTime":"2026-02-20T08:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:09 crc kubenswrapper[4948]: E0220 08:07:09.765393 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:09Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.770470 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.770532 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.770553 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.770577 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.770596 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:09Z","lastTransitionTime":"2026-02-20T08:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:09 crc kubenswrapper[4948]: E0220 08:07:09.788696 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:09Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.792182 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.792225 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.792236 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.792253 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.792269 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:09Z","lastTransitionTime":"2026-02-20T08:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:09 crc kubenswrapper[4948]: E0220 08:07:09.809631 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:09Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.814414 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.814458 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.814470 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.814488 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.814501 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:09Z","lastTransitionTime":"2026-02-20T08:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:09 crc kubenswrapper[4948]: E0220 08:07:09.826846 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:09Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.830199 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.830235 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.830243 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.830258 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.830270 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:09Z","lastTransitionTime":"2026-02-20T08:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:09 crc kubenswrapper[4948]: E0220 08:07:09.862490 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:09Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:09 crc kubenswrapper[4948]: E0220 08:07:09.862705 4948 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.867251 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.867302 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.867313 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.867328 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.867340 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:09Z","lastTransitionTime":"2026-02-20T08:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.970059 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.970097 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.970106 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.970119 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.970128 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:09Z","lastTransitionTime":"2026-02-20T08:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.998031 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 02:40:18.954398867 +0000 UTC Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.073378 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.073432 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.073449 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.073473 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.073491 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:10Z","lastTransitionTime":"2026-02-20T08:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.176997 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.177042 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.177056 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.177074 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.177086 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:10Z","lastTransitionTime":"2026-02-20T08:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.280101 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.280165 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.280182 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.280207 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.280224 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:10Z","lastTransitionTime":"2026-02-20T08:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.382888 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.382934 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.382946 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.382965 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.382996 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:10Z","lastTransitionTime":"2026-02-20T08:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.485900 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.486033 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.486067 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.486099 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.486124 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:10Z","lastTransitionTime":"2026-02-20T08:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.589320 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.589384 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.589401 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.589425 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.589443 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:10Z","lastTransitionTime":"2026-02-20T08:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.692284 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.692377 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.692397 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.692423 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.692446 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:10Z","lastTransitionTime":"2026-02-20T08:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.721883 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.721883 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.722000 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.722064 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:10 crc kubenswrapper[4948]: E0220 08:07:10.722253 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:10 crc kubenswrapper[4948]: E0220 08:07:10.722485 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:10 crc kubenswrapper[4948]: E0220 08:07:10.722570 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:10 crc kubenswrapper[4948]: E0220 08:07:10.722668 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.795590 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.795661 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.795754 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.795785 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.795807 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:10Z","lastTransitionTime":"2026-02-20T08:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.898673 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.898722 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.898737 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.898760 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.898777 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:10Z","lastTransitionTime":"2026-02-20T08:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.998207 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 05:08:17.046734525 +0000 UTC Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.002092 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.002158 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.002181 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.002214 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.002236 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:11Z","lastTransitionTime":"2026-02-20T08:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.105363 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.105427 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.105444 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.105471 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.105490 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:11Z","lastTransitionTime":"2026-02-20T08:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.209362 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.209428 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.209451 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.209479 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.209497 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:11Z","lastTransitionTime":"2026-02-20T08:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.312776 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.313270 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.313440 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.313634 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.313792 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:11Z","lastTransitionTime":"2026-02-20T08:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.416735 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.416800 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.416819 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.416845 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.416864 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:11Z","lastTransitionTime":"2026-02-20T08:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.520133 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.520220 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.520237 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.520262 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.520280 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:11Z","lastTransitionTime":"2026-02-20T08:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.624091 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.624158 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.624176 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.624200 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.624220 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:11Z","lastTransitionTime":"2026-02-20T08:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.728166 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.728232 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.728249 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.728276 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.728336 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:11Z","lastTransitionTime":"2026-02-20T08:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.739755 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ad7145-074b-49df-a0e9-cfec6e3974ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67b60ad2efa00639a153f39923bfdba3b40b6b3f5126207b4f0e05abdf9f43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:11Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.765039 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:11Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.780715 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:11Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.808450 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:11Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.827643 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:11Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.831578 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.831645 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.831667 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.831698 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.831721 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:11Z","lastTransitionTime":"2026-02-20T08:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.864591 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:44Z\\\",\\\"message\\\":\\\"ate admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z]\\\\nI0220 08:06:44.892233 6647 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/community-operators]} name:Service_openshift-marketplace/community-operators_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.189:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d389393c-7ba9-422c-b3f5-06e391d537d2}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:11Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.891499 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:11Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.918199 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ae14aa1-2ba5-4630-b0b9-97cef2fece9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://001bf218b5793a1101d9960b6d81173d67c5f6259c21d780623dad9808adfd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5f1450fc3f715e0082ad316af3f03ca0d2d8f32b5efa9850e0694864f79153e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a52eab83c2f040b6bd090f63f55a7da942f1b9a219ae29f14fc2b9930b07da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:11Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.934624 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.934714 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.934735 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.934759 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.934810 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:11Z","lastTransitionTime":"2026-02-20T08:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.937220 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:11Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.955069 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:11Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.978203 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:11Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.998458 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 11:11:43.074352609 +0000 UTC Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.000174 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:11Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.020893 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:12Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.038556 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.038607 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.038624 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.038650 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.038668 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:12Z","lastTransitionTime":"2026-02-20T08:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.048283 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f4093ec9d71181c58ad80a8c4ee3692148590df14aadaec96c0f54a843d1f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:07:02Z\\\",\\\"message\\\":\\\"2026-02-20T08:06:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b42b7938-8525-4811-9d9f-355344db9d45\\\\n2026-02-20T08:06:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b42b7938-8525-4811-9d9f-355344db9d45 to /host/opt/cni/bin/\\\\n2026-02-20T08:06:17Z [verbose] multus-daemon started\\\\n2026-02-20T08:06:17Z [verbose] Readiness Indicator file check\\\\n2026-02-20T08:07:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:12Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.081767 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:12Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.099284 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:12Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.113341 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:12Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.128061 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6cfa24bc2ee3daae226cdfb1ec830d8a97c86eed37ca8ce2d0c10878627bcc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cc8878e457802d4970c4db52fc332cf96d26ab80c90cbe1c3160b8c02b385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:12Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.141532 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xd86g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xd86g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:12Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.141772 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.141935 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.142233 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.142478 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.146022 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:12Z","lastTransitionTime":"2026-02-20T08:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.248776 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.249072 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.249081 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.249095 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.249105 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:12Z","lastTransitionTime":"2026-02-20T08:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.352356 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.352414 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.352428 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.352629 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.352641 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:12Z","lastTransitionTime":"2026-02-20T08:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.456632 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.456697 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.456719 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.456746 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.456766 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:12Z","lastTransitionTime":"2026-02-20T08:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.566713 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.566788 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.566810 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.566840 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.566864 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:12Z","lastTransitionTime":"2026-02-20T08:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.670149 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.670224 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.670246 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.670278 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.670301 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:12Z","lastTransitionTime":"2026-02-20T08:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.722339 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.722339 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.722364 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.722504 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:12 crc kubenswrapper[4948]: E0220 08:07:12.723280 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:12 crc kubenswrapper[4948]: E0220 08:07:12.723467 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:12 crc kubenswrapper[4948]: E0220 08:07:12.723652 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:12 crc kubenswrapper[4948]: E0220 08:07:12.723175 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.772595 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.773324 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.773675 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.773859 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.774030 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:12Z","lastTransitionTime":"2026-02-20T08:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.877107 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.877182 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.877204 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.877236 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.877254 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:12Z","lastTransitionTime":"2026-02-20T08:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.981924 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.982453 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.982603 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.982757 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.982885 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:12Z","lastTransitionTime":"2026-02-20T08:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.999713 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 13:33:41.867727325 +0000 UTC Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.086743 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.086781 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.086792 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.086813 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.086824 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:13Z","lastTransitionTime":"2026-02-20T08:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.189943 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.190404 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.190520 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.190637 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.190786 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:13Z","lastTransitionTime":"2026-02-20T08:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.294275 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.294327 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.294341 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.294361 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.294373 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:13Z","lastTransitionTime":"2026-02-20T08:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.398241 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.398314 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.398330 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.398363 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.398383 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:13Z","lastTransitionTime":"2026-02-20T08:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.500810 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.500852 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.500865 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.500886 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.500900 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:13Z","lastTransitionTime":"2026-02-20T08:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.604206 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.604279 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.604297 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.604325 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.604347 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:13Z","lastTransitionTime":"2026-02-20T08:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.707736 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.708207 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.708420 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.708590 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.708729 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:13Z","lastTransitionTime":"2026-02-20T08:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.722655 4948 scope.go:117] "RemoveContainer" containerID="bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.811254 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.811296 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.811311 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.811333 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.811348 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:13Z","lastTransitionTime":"2026-02-20T08:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.915187 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.915243 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.915257 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.915280 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.915294 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:13Z","lastTransitionTime":"2026-02-20T08:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.000653 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 18:00:13.65912416 +0000 UTC Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.018959 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.019022 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.019032 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.019049 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.019059 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:14Z","lastTransitionTime":"2026-02-20T08:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.122018 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.122084 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.122100 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.122125 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.122143 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:14Z","lastTransitionTime":"2026-02-20T08:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.224424 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.224459 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.224468 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.224482 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.224491 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:14Z","lastTransitionTime":"2026-02-20T08:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.312281 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovnkube-controller/2.log" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.315477 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerStarted","Data":"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b"} Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.316350 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.327803 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.327894 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.327921 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.327956 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.328036 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:14Z","lastTransitionTime":"2026-02-20T08:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.339659 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.356592 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.370180 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.394662 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:44Z\\\",\\\"message\\\":\\\"ate admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z]\\\\nI0220 08:06:44.892233 6647 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/community-operators]} name:Service_openshift-marketplace/community-operators_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.189:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d389393c-7ba9-422c-b3f5-06e391d537d2}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.407159 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ad7145-074b-49df-a0e9-cfec6e3974ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67b60ad2efa00639a153f39923bfdba3b40b6b3f5126207b4f0e05abdf9f43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.422032 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.431056 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.431086 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.431096 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.431110 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.431120 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:14Z","lastTransitionTime":"2026-02-20T08:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.437517 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.451512 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.465423 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.477552 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ae14aa1-2ba5-4630-b0b9-97cef2fece9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://001bf218b5793a1101d9960b6d81173d67c5f6259c21d780623dad9808adfd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5f1450fc3f715e0082ad316af3f03ca0d2d8f32b5efa9850e0694864f79153e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a52eab83c2f040b6bd090f63f55a7da942f1b9a219ae29f14fc2b9930b07da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.490217 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.504746 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f4093ec9d71181c58ad80a8c4ee3692148590df14aadaec96c0f54a843d1f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:07:02Z\\\",\\\"message\\\":\\\"2026-02-20T08:06:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b42b7938-8525-4811-9d9f-355344db9d45\\\\n2026-02-20T08:06:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b42b7938-8525-4811-9d9f-355344db9d45 to /host/opt/cni/bin/\\\\n2026-02-20T08:06:17Z [verbose] multus-daemon started\\\\n2026-02-20T08:06:17Z [verbose] Readiness Indicator file check\\\\n2026-02-20T08:07:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.520098 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.532231 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.533929 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.533956 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.533980 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.534036 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.534051 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:14Z","lastTransitionTime":"2026-02-20T08:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.548021 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.561532 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6cfa24bc2ee3daae226cdfb1ec830d8a97c86eed37ca8ce2d0c10878627bcc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cc8878e457802d4970c4db52fc332cf96d26ab80c90cbe1c3160b8c02b385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.575929 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xd86g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xd86g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.603114 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.618547 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.635817 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.635844 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.635853 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.635866 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.635875 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:14Z","lastTransitionTime":"2026-02-20T08:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.722375 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.722371 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:14 crc kubenswrapper[4948]: E0220 08:07:14.722562 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.722411 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:14 crc kubenswrapper[4948]: E0220 08:07:14.722721 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.722371 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:14 crc kubenswrapper[4948]: E0220 08:07:14.722842 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:14 crc kubenswrapper[4948]: E0220 08:07:14.722900 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.738495 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.738544 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.738560 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.738578 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.738595 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:14Z","lastTransitionTime":"2026-02-20T08:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.841866 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.841927 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.841948 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.842005 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.842027 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:14Z","lastTransitionTime":"2026-02-20T08:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.945632 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.945701 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.945719 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.945744 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.945764 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:14Z","lastTransitionTime":"2026-02-20T08:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.001407 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 21:27:53.104968457 +0000 UTC Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.048674 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.048749 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.048777 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.048807 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.048828 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:15Z","lastTransitionTime":"2026-02-20T08:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.150950 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.151029 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.151042 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.151065 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.151080 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:15Z","lastTransitionTime":"2026-02-20T08:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.253483 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.253517 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.253527 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.253545 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.253555 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:15Z","lastTransitionTime":"2026-02-20T08:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.322443 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovnkube-controller/3.log" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.323720 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovnkube-controller/2.log" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.327965 4948 generic.go:334] "Generic (PLEG): container finished" podID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerID="7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b" exitCode=1 Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.328072 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerDied","Data":"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b"} Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.328185 4948 scope.go:117] "RemoveContainer" containerID="bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.329370 4948 scope.go:117] "RemoveContainer" containerID="7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b" Feb 20 08:07:15 crc kubenswrapper[4948]: E0220 08:07:15.329730 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.349617 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.357483 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.357556 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.357574 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.357602 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.357620 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:15Z","lastTransitionTime":"2026-02-20T08:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.370279 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6cfa24bc2ee3daae226cdfb1ec830d8a97c86eed37ca8ce2d0c10878627bcc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cc8878e457802d4970c4db52fc332cf96d26ab80c90cbe1c3160b8c02b385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.390297 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xd86g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xd86g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.427300 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.451627 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.460571 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.460622 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.460639 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.460664 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.460686 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:15Z","lastTransitionTime":"2026-02-20T08:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.469553 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.485185 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.496872 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.519695 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:44Z\\\",\\\"message\\\":\\\"ate admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z]\\\\nI0220 08:06:44.892233 6647 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/community-operators]} name:Service_openshift-marketplace/community-operators_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.189:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d389393c-7ba9-422c-b3f5-06e391d537d2}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:07:14Z\\\",\\\"message\\\":\\\" (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0220 08:07:14.607038 7089 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0220 08:07:14.607659 7089 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0220 08:07:14.607737 7089 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:07:14.607794 7089 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0220 08:07:14.608304 7089 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0220 08:07:14.608367 7089 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0220 08:07:14.608418 7089 factory.go:656] Stopping watch factory\\\\nI0220 08:07:14.608455 7089 ovnkube.go:599] Stopped ovnkube\\\\nI0220 08:07:14.608456 7089 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.532708 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ad7145-074b-49df-a0e9-cfec6e3974ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67b60ad2efa00639a153f39923bfdba3b40b6b3f5126207b4f0e05abdf9f43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.546402 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.564524 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.564584 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.564598 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.564618 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.564635 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:15Z","lastTransitionTime":"2026-02-20T08:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.564992 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.579813 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.596609 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.611025 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ae14aa1-2ba5-4630-b0b9-97cef2fece9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://001bf218b5793a1101d9960b6d81173d67c5f6259c21d780623dad9808adfd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5f1450fc3f715e0082ad316af3f03ca0d2d8f32b5efa9850e0694864f79153e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a52eab83c2f040b6bd090f63f55a7da942f1b9a219ae29f14fc2b9930b07da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.626144 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.641477 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f4093ec9d71181c58ad80a8c4ee3692148590df14aadaec96c0f54a843d1f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:07:02Z\\\",\\\"message\\\":\\\"2026-02-20T08:06:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b42b7938-8525-4811-9d9f-355344db9d45\\\\n2026-02-20T08:06:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b42b7938-8525-4811-9d9f-355344db9d45 to /host/opt/cni/bin/\\\\n2026-02-20T08:06:17Z [verbose] multus-daemon started\\\\n2026-02-20T08:06:17Z [verbose] Readiness Indicator file check\\\\n2026-02-20T08:07:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.653913 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.668376 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.668408 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.668416 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.668431 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.668442 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:15Z","lastTransitionTime":"2026-02-20T08:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.668459 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.770172 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.770210 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.770221 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.770240 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.770251 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:15Z","lastTransitionTime":"2026-02-20T08:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.873110 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.873165 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.873178 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.873197 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.873209 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:15Z","lastTransitionTime":"2026-02-20T08:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.976174 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.976240 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.976253 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.976281 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.976300 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:15Z","lastTransitionTime":"2026-02-20T08:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:16 crc kubenswrapper[4948]: I0220 08:07:16.001856 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 06:39:37.670922273 +0000 UTC Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:16.520379 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:16.520436 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:16.520576 4948 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:16.520642 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 08:08:20.520621404 +0000 UTC m=+149.495116224 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:16.520844 4948 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:16.520869 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 08:08:20.52086201 +0000 UTC m=+149.495356820 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:16.621404 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:16.621472 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:16.621520 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:16.621630 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:16.621645 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:16.621658 4948 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:16.621694 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-20 08:08:20.621681948 +0000 UTC m=+149.596176768 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:16.621872 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:20.621863943 +0000 UTC m=+149.596358763 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:16.621923 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:16.621934 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:16.621942 4948 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:16.621991 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-20 08:08:20.621983595 +0000 UTC m=+149.596478415 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.002356 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 04:32:58.656160035 +0000 UTC Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.282704 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:17.282850 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.282854 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.282881 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.283128 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:17.282947 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:17.283477 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:17.283860 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.285536 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.285573 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.285585 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.285604 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.285618 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:17Z","lastTransitionTime":"2026-02-20T08:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.289050 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovnkube-controller/3.log" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.293252 4948 scope.go:117] "RemoveContainer" containerID="7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b" Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:17.293886 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.311941 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.328096 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ae14aa1-2ba5-4630-b0b9-97cef2fece9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://001bf218b5793a1101d9960b6d81173d67c5f6259c21d780623dad9808adfd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5f1450fc3f715e0082ad316af3f03ca0d2d8f32b5efa9850e0694864f79153e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a52eab83c2f040b6bd090f63f55a7da942f1b9a219ae29f14fc2b9930b07da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.347424 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.364076 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.379556 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.388073 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.388119 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.388128 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.388144 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.388157 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:17Z","lastTransitionTime":"2026-02-20T08:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.398861 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.412584 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.430621 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f4093ec9d71181c58ad80a8c4ee3692148590df14aadaec96c0f54a843d1f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:07:02Z\\\",\\\"message\\\":\\\"2026-02-20T08:06:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b42b7938-8525-4811-9d9f-355344db9d45\\\\n2026-02-20T08:06:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b42b7938-8525-4811-9d9f-355344db9d45 to /host/opt/cni/bin/\\\\n2026-02-20T08:06:17Z [verbose] multus-daemon started\\\\n2026-02-20T08:06:17Z [verbose] Readiness Indicator file check\\\\n2026-02-20T08:07:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.443564 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xd86g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xd86g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.466946 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.482187 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.491265 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.491337 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.491359 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.491391 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.491412 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:17Z","lastTransitionTime":"2026-02-20T08:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.497547 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.515041 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6cfa24bc2ee3daae226cdfb1ec830d8a97c86eed37ca8ce2d0c10878627bcc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cc8878e457802d4970c4db52fc332cf96d26ab80c90cbe1c3160b8c02b385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.531689 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.554760 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:07:14Z\\\",\\\"message\\\":\\\" (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0220 08:07:14.607038 7089 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0220 08:07:14.607659 7089 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0220 08:07:14.607737 7089 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:07:14.607794 7089 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0220 08:07:14.608304 7089 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0220 08:07:14.608367 7089 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0220 08:07:14.608418 7089 factory.go:656] Stopping watch factory\\\\nI0220 08:07:14.608455 7089 ovnkube.go:599] Stopped ovnkube\\\\nI0220 08:07:14.608456 7089 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:07:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.569598 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ad7145-074b-49df-a0e9-cfec6e3974ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67b60ad2efa00639a153f39923bfdba3b40b6b3f5126207b4f0e05abdf9f43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.587621 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.593830 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.593893 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.593913 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.593938 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.593958 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:17Z","lastTransitionTime":"2026-02-20T08:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.606965 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.627021 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.697024 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.697087 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.697107 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.697130 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.697145 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:17Z","lastTransitionTime":"2026-02-20T08:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.800106 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.800403 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.800473 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.800537 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.800596 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:17Z","lastTransitionTime":"2026-02-20T08:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.904109 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.904175 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.904198 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.904228 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.904251 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:17Z","lastTransitionTime":"2026-02-20T08:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.003280 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 11:08:55.305869874 +0000 UTC Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.008403 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.008462 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.008479 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.008505 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.008522 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:18Z","lastTransitionTime":"2026-02-20T08:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.111604 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.111669 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.111690 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.111718 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.111741 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:18Z","lastTransitionTime":"2026-02-20T08:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.214915 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.215028 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.215066 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.215097 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.215117 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:18Z","lastTransitionTime":"2026-02-20T08:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.318431 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.318484 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.318498 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.318518 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.318532 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:18Z","lastTransitionTime":"2026-02-20T08:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.422310 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.422652 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.422666 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.422682 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.422693 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:18Z","lastTransitionTime":"2026-02-20T08:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.525229 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.525299 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.525317 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.525342 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.525361 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:18Z","lastTransitionTime":"2026-02-20T08:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.628820 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.629056 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.629085 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.629120 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.629144 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:18Z","lastTransitionTime":"2026-02-20T08:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.722426 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.722498 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.722424 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:18 crc kubenswrapper[4948]: E0220 08:07:18.722686 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:18 crc kubenswrapper[4948]: E0220 08:07:18.722791 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:18 crc kubenswrapper[4948]: E0220 08:07:18.722891 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.731698 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.731765 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.731787 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.731815 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.731840 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:18Z","lastTransitionTime":"2026-02-20T08:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.835804 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.835862 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.835879 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.835901 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.835918 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:18Z","lastTransitionTime":"2026-02-20T08:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.939032 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.939091 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.939113 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.939140 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.939163 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:18Z","lastTransitionTime":"2026-02-20T08:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.004556 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 01:32:47.679988356 +0000 UTC Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.041486 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.041609 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.041628 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.041653 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.041671 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:19Z","lastTransitionTime":"2026-02-20T08:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.144544 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.144612 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.144635 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.144659 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.144677 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:19Z","lastTransitionTime":"2026-02-20T08:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.248090 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.248174 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.248205 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.248248 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.248277 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:19Z","lastTransitionTime":"2026-02-20T08:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.351029 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.351091 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.351114 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.351144 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.351166 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:19Z","lastTransitionTime":"2026-02-20T08:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.454125 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.454184 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.454202 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.454237 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.454260 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:19Z","lastTransitionTime":"2026-02-20T08:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.557886 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.558008 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.558032 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.558061 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.558082 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:19Z","lastTransitionTime":"2026-02-20T08:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.660930 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.661008 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.661028 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.661055 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.661074 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:19Z","lastTransitionTime":"2026-02-20T08:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.721577 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:19 crc kubenswrapper[4948]: E0220 08:07:19.721771 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.763837 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.763894 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.763912 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.763934 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.763951 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:19Z","lastTransitionTime":"2026-02-20T08:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.866763 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.866815 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.866832 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.866853 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.866870 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:19Z","lastTransitionTime":"2026-02-20T08:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.969936 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.970021 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.970046 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.970072 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.970096 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:19Z","lastTransitionTime":"2026-02-20T08:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.004871 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 07:59:29.364927949 +0000 UTC Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.073631 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.073701 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.073728 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.073758 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.073782 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:20Z","lastTransitionTime":"2026-02-20T08:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.083131 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.083395 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.083441 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.083478 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.083496 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:20Z","lastTransitionTime":"2026-02-20T08:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:20 crc kubenswrapper[4948]: E0220 08:07:20.105578 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.110568 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.110629 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.110720 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.110751 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.110768 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:20Z","lastTransitionTime":"2026-02-20T08:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:20 crc kubenswrapper[4948]: E0220 08:07:20.132568 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.142525 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.143119 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.143151 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.143181 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.143202 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:20Z","lastTransitionTime":"2026-02-20T08:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:20 crc kubenswrapper[4948]: E0220 08:07:20.165532 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.171558 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.171622 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.171641 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.171668 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.171685 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:20Z","lastTransitionTime":"2026-02-20T08:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:20 crc kubenswrapper[4948]: E0220 08:07:20.193911 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.198702 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.198756 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.198774 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.198800 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.198820 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:20Z","lastTransitionTime":"2026-02-20T08:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:20 crc kubenswrapper[4948]: E0220 08:07:20.220357 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:20 crc kubenswrapper[4948]: E0220 08:07:20.220644 4948 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.223002 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.223069 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.223090 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.223113 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.223130 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:20Z","lastTransitionTime":"2026-02-20T08:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.326882 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.326948 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.327014 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.327048 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.327071 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:20Z","lastTransitionTime":"2026-02-20T08:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.429731 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.429787 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.429811 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.429839 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.429860 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:20Z","lastTransitionTime":"2026-02-20T08:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.536112 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.536160 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.536176 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.536194 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.536205 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:20Z","lastTransitionTime":"2026-02-20T08:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.639927 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.640024 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.640046 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.640072 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.640089 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:20Z","lastTransitionTime":"2026-02-20T08:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.721773 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.721847 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.721773 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:20 crc kubenswrapper[4948]: E0220 08:07:20.722018 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:20 crc kubenswrapper[4948]: E0220 08:07:20.722164 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:20 crc kubenswrapper[4948]: E0220 08:07:20.722303 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.742697 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.742752 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.742790 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.742821 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.742844 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:20Z","lastTransitionTime":"2026-02-20T08:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.845555 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.845614 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.845637 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.845660 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.845681 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:20Z","lastTransitionTime":"2026-02-20T08:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.948848 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.948902 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.948912 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.948928 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.948942 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:20Z","lastTransitionTime":"2026-02-20T08:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.006458 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 23:15:07.587189321 +0000 UTC Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.052377 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.052426 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.052445 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.052467 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.052515 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:21Z","lastTransitionTime":"2026-02-20T08:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.156905 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.156952 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.156961 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.157003 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.157016 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:21Z","lastTransitionTime":"2026-02-20T08:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.259757 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.259801 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.259813 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.259909 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.259927 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:21Z","lastTransitionTime":"2026-02-20T08:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.363406 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.363447 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.363456 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.363472 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.363485 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:21Z","lastTransitionTime":"2026-02-20T08:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.465724 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.465752 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.465761 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.465774 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.465783 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:21Z","lastTransitionTime":"2026-02-20T08:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.567944 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.567992 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.568001 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.568020 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.568031 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:21Z","lastTransitionTime":"2026-02-20T08:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.671858 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.671928 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.671941 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.671964 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.671999 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:21Z","lastTransitionTime":"2026-02-20T08:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.722684 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:21 crc kubenswrapper[4948]: E0220 08:07:21.722952 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.746097 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.765608 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ae14aa1-2ba5-4630-b0b9-97cef2fece9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://001bf218b5793a1101d9960b6d81173d67c5f6259c21d780623dad9808adfd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5f1450fc3f715e0082ad316af3f03ca0d2d8f32b5efa9850e0694864f79153e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a52eab83c2f040b6bd090f63f55a7da942f1b9a219ae29f14fc2b9930b07da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.774264 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.774312 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.774327 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.774352 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.774367 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:21Z","lastTransitionTime":"2026-02-20T08:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.779237 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.795000 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.812402 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.826372 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.844483 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.864132 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f4093ec9d71181c58ad80a8c4ee3692148590df14aadaec96c0f54a843d1f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:07:02Z\\\",\\\"message\\\":\\\"2026-02-20T08:06:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b42b7938-8525-4811-9d9f-355344db9d45\\\\n2026-02-20T08:06:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b42b7938-8525-4811-9d9f-355344db9d45 to /host/opt/cni/bin/\\\\n2026-02-20T08:06:17Z [verbose] multus-daemon started\\\\n2026-02-20T08:06:17Z [verbose] Readiness Indicator file check\\\\n2026-02-20T08:07:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.876634 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.876674 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.876686 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.876702 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.876715 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:21Z","lastTransitionTime":"2026-02-20T08:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.881407 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xd86g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xd86g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.902329 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.915541 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.928083 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.942271 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6cfa24bc2ee3daae226cdfb1ec830d8a97c86eed37ca8ce2d0c10878627bcc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cc8878e457802d4970c4db52fc332cf96d26ab80c90cbe1c3160b8c02b385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.955130 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.973719 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:07:14Z\\\",\\\"message\\\":\\\" (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0220 08:07:14.607038 7089 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0220 08:07:14.607659 7089 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0220 08:07:14.607737 7089 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:07:14.607794 7089 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0220 08:07:14.608304 7089 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0220 08:07:14.608367 7089 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0220 08:07:14.608418 7089 factory.go:656] Stopping watch factory\\\\nI0220 08:07:14.608455 7089 ovnkube.go:599] Stopped ovnkube\\\\nI0220 08:07:14.608456 7089 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:07:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.978890 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.978922 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.978934 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.978952 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.978963 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:21Z","lastTransitionTime":"2026-02-20T08:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.988587 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ad7145-074b-49df-a0e9-cfec6e3974ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67b60ad2efa00639a153f39923bfdba3b40b6b3f5126207b4f0e05abdf9f43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.006839 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 15:28:44.025217868 +0000 UTC Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.007312 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.020550 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.038524 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.082032 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.082091 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.082102 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.082121 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.082134 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:22Z","lastTransitionTime":"2026-02-20T08:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.185067 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.185151 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.185175 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.185207 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.185233 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:22Z","lastTransitionTime":"2026-02-20T08:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.287862 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.287942 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.288025 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.288080 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.288104 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:22Z","lastTransitionTime":"2026-02-20T08:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.391176 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.391251 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.391270 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.391294 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.391313 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:22Z","lastTransitionTime":"2026-02-20T08:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.494629 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.494690 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.494707 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.494732 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.494750 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:22Z","lastTransitionTime":"2026-02-20T08:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.598051 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.598109 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.598145 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.598169 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.598187 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:22Z","lastTransitionTime":"2026-02-20T08:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.701073 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.701132 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.701152 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.701176 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.701196 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:22Z","lastTransitionTime":"2026-02-20T08:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.721818 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.721944 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:22 crc kubenswrapper[4948]: E0220 08:07:22.722036 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.721944 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:22 crc kubenswrapper[4948]: E0220 08:07:22.722202 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:22 crc kubenswrapper[4948]: E0220 08:07:22.722309 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.804817 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.804879 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.804900 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.804926 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.804947 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:22Z","lastTransitionTime":"2026-02-20T08:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.908448 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.908489 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.908500 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.908515 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.908527 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:22Z","lastTransitionTime":"2026-02-20T08:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.008005 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 13:51:16.041023369 +0000 UTC Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.011707 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.011771 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.011789 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.011817 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.011836 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:23Z","lastTransitionTime":"2026-02-20T08:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.115132 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.115197 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.115221 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.115251 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.115276 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:23Z","lastTransitionTime":"2026-02-20T08:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.219147 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.219236 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.219264 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.219298 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.219323 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:23Z","lastTransitionTime":"2026-02-20T08:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.321756 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.321833 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.321859 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.321891 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.321912 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:23Z","lastTransitionTime":"2026-02-20T08:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.444116 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.444210 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.444228 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.444259 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.444283 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:23Z","lastTransitionTime":"2026-02-20T08:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.548780 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.548861 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.548888 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.548935 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.548960 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:23Z","lastTransitionTime":"2026-02-20T08:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.652595 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.652851 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.653029 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.653236 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.653420 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:23Z","lastTransitionTime":"2026-02-20T08:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.722524 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:23 crc kubenswrapper[4948]: E0220 08:07:23.722879 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.756827 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.756926 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.756945 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.756994 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.757009 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:23Z","lastTransitionTime":"2026-02-20T08:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.860730 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.860930 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.861196 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.861393 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.861547 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:23Z","lastTransitionTime":"2026-02-20T08:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.964820 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.964893 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.964922 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.964958 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.965017 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:23Z","lastTransitionTime":"2026-02-20T08:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.008914 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 13:55:27.866622145 +0000 UTC Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.069127 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.069193 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.069216 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.069245 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.069265 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:24Z","lastTransitionTime":"2026-02-20T08:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.172533 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.172599 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.172625 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.172652 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.172671 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:24Z","lastTransitionTime":"2026-02-20T08:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.274551 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.274604 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.274621 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.274644 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.274660 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:24Z","lastTransitionTime":"2026-02-20T08:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.377527 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.377586 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.377611 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.377646 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.377671 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:24Z","lastTransitionTime":"2026-02-20T08:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.480469 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.480526 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.480543 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.480570 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.480589 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:24Z","lastTransitionTime":"2026-02-20T08:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.584518 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.585544 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.585769 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.585942 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.586131 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:24Z","lastTransitionTime":"2026-02-20T08:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.722737 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.723165 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:24 crc kubenswrapper[4948]: E0220 08:07:24.723316 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.723405 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:24 crc kubenswrapper[4948]: E0220 08:07:24.723498 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:24 crc kubenswrapper[4948]: E0220 08:07:24.723644 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.723699 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.723758 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.723778 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.723801 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.723819 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:24Z","lastTransitionTime":"2026-02-20T08:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.826173 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.826254 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.826273 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.826305 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.826394 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:24Z","lastTransitionTime":"2026-02-20T08:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.929923 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.930038 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.930056 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.930081 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.930102 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:24Z","lastTransitionTime":"2026-02-20T08:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.010053 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 23:00:32.684727381 +0000 UTC Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.032744 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.032793 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.032808 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.032831 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.032846 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:25Z","lastTransitionTime":"2026-02-20T08:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.135885 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.136002 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.136023 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.136048 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.136067 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:25Z","lastTransitionTime":"2026-02-20T08:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.239253 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.239317 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.239334 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.239362 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.239380 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:25Z","lastTransitionTime":"2026-02-20T08:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.342483 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.342538 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.342557 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.342583 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.342601 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:25Z","lastTransitionTime":"2026-02-20T08:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.445040 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.445101 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.445143 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.445171 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.445190 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:25Z","lastTransitionTime":"2026-02-20T08:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.547512 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.547548 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.547556 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.547569 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.547579 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:25Z","lastTransitionTime":"2026-02-20T08:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.650624 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.650665 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.650704 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.650722 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.650735 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:25Z","lastTransitionTime":"2026-02-20T08:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.721843 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:25 crc kubenswrapper[4948]: E0220 08:07:25.722012 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.752620 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.753482 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.753543 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.753580 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.753604 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:25Z","lastTransitionTime":"2026-02-20T08:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.856883 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.856945 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.856967 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.857020 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.857043 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:25Z","lastTransitionTime":"2026-02-20T08:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.959365 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.959437 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.959461 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.959493 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.959517 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:25Z","lastTransitionTime":"2026-02-20T08:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.010210 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 18:22:37.494920974 +0000 UTC Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.061836 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.061867 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.061876 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.061889 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.061897 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:26Z","lastTransitionTime":"2026-02-20T08:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.164348 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.164452 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.164525 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.164557 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.164580 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:26Z","lastTransitionTime":"2026-02-20T08:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.267176 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.267509 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.267619 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.267726 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.267825 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:26Z","lastTransitionTime":"2026-02-20T08:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.371204 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.371270 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.371288 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.371315 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.371346 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:26Z","lastTransitionTime":"2026-02-20T08:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.474735 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.475115 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.475210 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.475291 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.475359 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:26Z","lastTransitionTime":"2026-02-20T08:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.578707 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.579079 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.579233 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.579383 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.579523 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:26Z","lastTransitionTime":"2026-02-20T08:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.682760 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.683607 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.683747 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.683880 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.684097 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:26Z","lastTransitionTime":"2026-02-20T08:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.722149 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.722198 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:26 crc kubenswrapper[4948]: E0220 08:07:26.722389 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:26 crc kubenswrapper[4948]: E0220 08:07:26.722524 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.722717 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:26 crc kubenswrapper[4948]: E0220 08:07:26.723098 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.786754 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.786791 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.786802 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.786816 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.786827 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:26Z","lastTransitionTime":"2026-02-20T08:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.889040 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.889093 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.889105 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.889119 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.889131 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:26Z","lastTransitionTime":"2026-02-20T08:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.992032 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.992112 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.992138 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.992167 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.992186 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:26Z","lastTransitionTime":"2026-02-20T08:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.011393 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 12:04:10.677207005 +0000 UTC Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.095634 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.096045 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.096077 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.096110 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.096135 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:27Z","lastTransitionTime":"2026-02-20T08:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.199403 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.199454 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.199466 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.199486 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.199504 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:27Z","lastTransitionTime":"2026-02-20T08:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.303033 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.303121 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.303140 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.303168 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.303187 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:27Z","lastTransitionTime":"2026-02-20T08:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.406599 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.406949 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.407248 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.407455 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.407647 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:27Z","lastTransitionTime":"2026-02-20T08:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.511269 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.511631 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.511786 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.511934 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.512170 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:27Z","lastTransitionTime":"2026-02-20T08:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.617613 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.618016 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.618244 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.618431 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.618634 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:27Z","lastTransitionTime":"2026-02-20T08:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.722862 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:27 crc kubenswrapper[4948]: E0220 08:07:27.724003 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.724376 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.724499 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.724582 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.724671 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.724761 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:27Z","lastTransitionTime":"2026-02-20T08:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.726311 4948 scope.go:117] "RemoveContainer" containerID="7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b" Feb 20 08:07:27 crc kubenswrapper[4948]: E0220 08:07:27.726606 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.828044 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.828120 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.828138 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.828167 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.828188 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:27Z","lastTransitionTime":"2026-02-20T08:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.931828 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.932269 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.932495 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.932698 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.932859 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:27Z","lastTransitionTime":"2026-02-20T08:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.012020 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 02:33:57.306471255 +0000 UTC Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.036324 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.036387 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.036402 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.036421 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.036437 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:28Z","lastTransitionTime":"2026-02-20T08:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.140100 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.140511 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.140681 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.140845 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.141033 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:28Z","lastTransitionTime":"2026-02-20T08:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.244342 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.244414 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.244454 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.244492 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.244520 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:28Z","lastTransitionTime":"2026-02-20T08:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.347597 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.347665 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.347688 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.347722 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.347743 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:28Z","lastTransitionTime":"2026-02-20T08:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.451582 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.452291 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.452537 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.452753 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.452925 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:28Z","lastTransitionTime":"2026-02-20T08:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.555671 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.556115 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.556380 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.556597 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.556808 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:28Z","lastTransitionTime":"2026-02-20T08:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.659458 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.659724 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.659795 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.660049 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.660129 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:28Z","lastTransitionTime":"2026-02-20T08:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.721985 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.722094 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.722046 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:28 crc kubenswrapper[4948]: E0220 08:07:28.722402 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:28 crc kubenswrapper[4948]: E0220 08:07:28.722561 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:28 crc kubenswrapper[4948]: E0220 08:07:28.722694 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.762251 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.762281 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.762290 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.762309 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.762319 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:28Z","lastTransitionTime":"2026-02-20T08:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.865162 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.865587 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.865609 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.865634 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.865651 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:28Z","lastTransitionTime":"2026-02-20T08:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.968564 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.968616 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.968628 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.968645 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.968659 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:28Z","lastTransitionTime":"2026-02-20T08:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.012855 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 08:38:50.010146446 +0000 UTC Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.072239 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.072281 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.072292 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.072309 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.072321 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:29Z","lastTransitionTime":"2026-02-20T08:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.175541 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.175621 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.175639 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.175668 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.175690 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:29Z","lastTransitionTime":"2026-02-20T08:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.279023 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.279092 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.279116 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.279142 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.279165 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:29Z","lastTransitionTime":"2026-02-20T08:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.382108 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.382168 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.382185 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.382207 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.382226 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:29Z","lastTransitionTime":"2026-02-20T08:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.485138 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.485206 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.485223 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.485252 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.485271 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:29Z","lastTransitionTime":"2026-02-20T08:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.587834 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.587892 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.587909 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.587932 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.587949 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:29Z","lastTransitionTime":"2026-02-20T08:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.690841 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.690900 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.690910 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.690928 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.690941 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:29Z","lastTransitionTime":"2026-02-20T08:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.722417 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:29 crc kubenswrapper[4948]: E0220 08:07:29.722965 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.794300 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.794345 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.794357 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.794375 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.794387 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:29Z","lastTransitionTime":"2026-02-20T08:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.896844 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.896911 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.896930 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.896954 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.897002 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:29Z","lastTransitionTime":"2026-02-20T08:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.000363 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.000425 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.000443 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.000466 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.000486 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:30Z","lastTransitionTime":"2026-02-20T08:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.013786 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 07:03:19.413260893 +0000 UTC Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.104024 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.104084 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.104101 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.104122 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.104136 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:30Z","lastTransitionTime":"2026-02-20T08:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.206529 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.206569 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.206581 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.206598 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.206612 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:30Z","lastTransitionTime":"2026-02-20T08:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.310257 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.310314 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.310330 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.310355 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.310372 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:30Z","lastTransitionTime":"2026-02-20T08:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.387942 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.388033 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.388052 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.388080 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.388102 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:30Z","lastTransitionTime":"2026-02-20T08:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.447898 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64"] Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.448377 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.452017 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.453364 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.459705 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.459734 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.484688 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=77.484657787 podStartE2EDuration="1m17.484657787s" podCreationTimestamp="2026-02-20 08:06:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:07:30.484400241 +0000 UTC m=+99.458895081" watchObservedRunningTime="2026-02-20 08:07:30.484657787 +0000 UTC m=+99.459152637" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.486960 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/1e7d2472-cdda-4f5b-996f-547bb51497d8-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-nkq64\" (UID: \"1e7d2472-cdda-4f5b-996f-547bb51497d8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.487107 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1e7d2472-cdda-4f5b-996f-547bb51497d8-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-nkq64\" (UID: \"1e7d2472-cdda-4f5b-996f-547bb51497d8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.487169 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/1e7d2472-cdda-4f5b-996f-547bb51497d8-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-nkq64\" (UID: \"1e7d2472-cdda-4f5b-996f-547bb51497d8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.487259 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1e7d2472-cdda-4f5b-996f-547bb51497d8-service-ca\") pod \"cluster-version-operator-5c965bbfc6-nkq64\" (UID: \"1e7d2472-cdda-4f5b-996f-547bb51497d8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.487304 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e7d2472-cdda-4f5b-996f-547bb51497d8-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-nkq64\" (UID: \"1e7d2472-cdda-4f5b-996f-547bb51497d8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.532262 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-x7s56" podStartSLOduration=76.532237523 podStartE2EDuration="1m16.532237523s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:07:30.532015438 +0000 UTC m=+99.506510258" watchObservedRunningTime="2026-02-20 08:07:30.532237523 +0000 UTC m=+99.506732383" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.588080 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1e7d2472-cdda-4f5b-996f-547bb51497d8-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-nkq64\" (UID: \"1e7d2472-cdda-4f5b-996f-547bb51497d8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.588128 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/1e7d2472-cdda-4f5b-996f-547bb51497d8-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-nkq64\" (UID: \"1e7d2472-cdda-4f5b-996f-547bb51497d8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.588088 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=78.588053027 podStartE2EDuration="1m18.588053027s" podCreationTimestamp="2026-02-20 08:06:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:07:30.56741853 +0000 UTC m=+99.541913350" watchObservedRunningTime="2026-02-20 08:07:30.588053027 +0000 UTC m=+99.562547887" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.588879 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1e7d2472-cdda-4f5b-996f-547bb51497d8-service-ca\") pod \"cluster-version-operator-5c965bbfc6-nkq64\" (UID: \"1e7d2472-cdda-4f5b-996f-547bb51497d8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.589126 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/1e7d2472-cdda-4f5b-996f-547bb51497d8-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-nkq64\" (UID: \"1e7d2472-cdda-4f5b-996f-547bb51497d8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.588162 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1e7d2472-cdda-4f5b-996f-547bb51497d8-service-ca\") pod \"cluster-version-operator-5c965bbfc6-nkq64\" (UID: \"1e7d2472-cdda-4f5b-996f-547bb51497d8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.589278 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e7d2472-cdda-4f5b-996f-547bb51497d8-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-nkq64\" (UID: \"1e7d2472-cdda-4f5b-996f-547bb51497d8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.589327 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/1e7d2472-cdda-4f5b-996f-547bb51497d8-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-nkq64\" (UID: \"1e7d2472-cdda-4f5b-996f-547bb51497d8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.589367 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/1e7d2472-cdda-4f5b-996f-547bb51497d8-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-nkq64\" (UID: \"1e7d2472-cdda-4f5b-996f-547bb51497d8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.597157 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e7d2472-cdda-4f5b-996f-547bb51497d8-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-nkq64\" (UID: \"1e7d2472-cdda-4f5b-996f-547bb51497d8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.602029 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-nsk7l" podStartSLOduration=76.602004833 podStartE2EDuration="1m16.602004833s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:07:30.601757447 +0000 UTC m=+99.576252267" watchObservedRunningTime="2026-02-20 08:07:30.602004833 +0000 UTC m=+99.576499653" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.615329 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1e7d2472-cdda-4f5b-996f-547bb51497d8-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-nkq64\" (UID: \"1e7d2472-cdda-4f5b-996f-547bb51497d8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.635480 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" podStartSLOduration=76.635459219 podStartE2EDuration="1m16.635459219s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:07:30.617414964 +0000 UTC m=+99.591909834" watchObservedRunningTime="2026-02-20 08:07:30.635459219 +0000 UTC m=+99.609954039" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.667494 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=32.667471299 podStartE2EDuration="32.667471299s" podCreationTimestamp="2026-02-20 08:06:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:07:30.652210552 +0000 UTC m=+99.626705372" watchObservedRunningTime="2026-02-20 08:07:30.667471299 +0000 UTC m=+99.641966119" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.681735 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-z8rrb" podStartSLOduration=76.681705582 podStartE2EDuration="1m16.681705582s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:07:30.681022425 +0000 UTC m=+99.655517245" watchObservedRunningTime="2026-02-20 08:07:30.681705582 +0000 UTC m=+99.656200422" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.703150 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-8frp4" podStartSLOduration=76.703133738 podStartE2EDuration="1m16.703133738s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:07:30.701834947 +0000 UTC m=+99.676329777" watchObservedRunningTime="2026-02-20 08:07:30.703133738 +0000 UTC m=+99.677628558" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.716825 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podStartSLOduration=76.716807847 podStartE2EDuration="1m16.716807847s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:07:30.716351616 +0000 UTC m=+99.690846436" watchObservedRunningTime="2026-02-20 08:07:30.716807847 +0000 UTC m=+99.691302667" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.721689 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.721727 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.721768 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:30 crc kubenswrapper[4948]: E0220 08:07:30.721800 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:30 crc kubenswrapper[4948]: E0220 08:07:30.721907 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:30 crc kubenswrapper[4948]: E0220 08:07:30.722052 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.774383 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=78.774363714 podStartE2EDuration="1m18.774363714s" podCreationTimestamp="2026-02-20 08:06:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:07:30.773096403 +0000 UTC m=+99.747591223" watchObservedRunningTime="2026-02-20 08:07:30.774363714 +0000 UTC m=+99.748858534" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.775080 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.816368 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=45.816341115 podStartE2EDuration="45.816341115s" podCreationTimestamp="2026-02-20 08:06:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:07:30.799410067 +0000 UTC m=+99.773904887" watchObservedRunningTime="2026-02-20 08:07:30.816341115 +0000 UTC m=+99.790835935" Feb 20 08:07:31 crc kubenswrapper[4948]: I0220 08:07:31.014256 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 12:54:48.8422828 +0000 UTC Feb 20 08:07:31 crc kubenswrapper[4948]: I0220 08:07:31.014343 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Feb 20 08:07:31 crc kubenswrapper[4948]: I0220 08:07:31.024557 4948 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 20 08:07:31 crc kubenswrapper[4948]: I0220 08:07:31.388994 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" event={"ID":"1e7d2472-cdda-4f5b-996f-547bb51497d8","Type":"ContainerStarted","Data":"a4f04e4cb0bf2b7c7d9aef9c44aae1318cc8087599dd94e6f0ca316bcaec97db"} Feb 20 08:07:31 crc kubenswrapper[4948]: I0220 08:07:31.389082 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" event={"ID":"1e7d2472-cdda-4f5b-996f-547bb51497d8","Type":"ContainerStarted","Data":"fcd683dea6851584660c8afd3237ccb1cbd4919e2f27216d734520806e1173af"} Feb 20 08:07:31 crc kubenswrapper[4948]: I0220 08:07:31.414762 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" podStartSLOduration=77.414738207 podStartE2EDuration="1m17.414738207s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:07:31.414469371 +0000 UTC m=+100.388964211" watchObservedRunningTime="2026-02-20 08:07:31.414738207 +0000 UTC m=+100.389233047" Feb 20 08:07:31 crc kubenswrapper[4948]: I0220 08:07:31.722141 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:31 crc kubenswrapper[4948]: E0220 08:07:31.724589 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:32 crc kubenswrapper[4948]: I0220 08:07:32.712187 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs\") pod \"network-metrics-daemon-xd86g\" (UID: \"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\") " pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:32 crc kubenswrapper[4948]: E0220 08:07:32.712447 4948 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 08:07:32 crc kubenswrapper[4948]: E0220 08:07:32.713766 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs podName:62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd nodeName:}" failed. No retries permitted until 2026-02-20 08:08:36.713414616 +0000 UTC m=+165.687909476 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs") pod "network-metrics-daemon-xd86g" (UID: "62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 08:07:32 crc kubenswrapper[4948]: I0220 08:07:32.722246 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:32 crc kubenswrapper[4948]: I0220 08:07:32.722246 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:32 crc kubenswrapper[4948]: E0220 08:07:32.722453 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:32 crc kubenswrapper[4948]: E0220 08:07:32.722595 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:32 crc kubenswrapper[4948]: I0220 08:07:32.722880 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:32 crc kubenswrapper[4948]: E0220 08:07:32.723347 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:33 crc kubenswrapper[4948]: I0220 08:07:33.722182 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:33 crc kubenswrapper[4948]: E0220 08:07:33.722378 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:34 crc kubenswrapper[4948]: I0220 08:07:34.722375 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:34 crc kubenswrapper[4948]: I0220 08:07:34.722445 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:34 crc kubenswrapper[4948]: I0220 08:07:34.722503 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:34 crc kubenswrapper[4948]: E0220 08:07:34.722615 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:34 crc kubenswrapper[4948]: E0220 08:07:34.723023 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:34 crc kubenswrapper[4948]: E0220 08:07:34.723099 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:35 crc kubenswrapper[4948]: I0220 08:07:35.721714 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:35 crc kubenswrapper[4948]: E0220 08:07:35.721968 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:36 crc kubenswrapper[4948]: I0220 08:07:36.721988 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:36 crc kubenswrapper[4948]: I0220 08:07:36.722010 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:36 crc kubenswrapper[4948]: E0220 08:07:36.722135 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:36 crc kubenswrapper[4948]: I0220 08:07:36.722207 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:36 crc kubenswrapper[4948]: E0220 08:07:36.722314 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:36 crc kubenswrapper[4948]: E0220 08:07:36.722443 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:37 crc kubenswrapper[4948]: I0220 08:07:37.722112 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:37 crc kubenswrapper[4948]: E0220 08:07:37.722339 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:38 crc kubenswrapper[4948]: I0220 08:07:38.721464 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:38 crc kubenswrapper[4948]: I0220 08:07:38.721537 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:38 crc kubenswrapper[4948]: I0220 08:07:38.721482 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:38 crc kubenswrapper[4948]: E0220 08:07:38.721668 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:38 crc kubenswrapper[4948]: E0220 08:07:38.721792 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:38 crc kubenswrapper[4948]: E0220 08:07:38.721937 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:39 crc kubenswrapper[4948]: I0220 08:07:39.721950 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:39 crc kubenswrapper[4948]: E0220 08:07:39.722427 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:40 crc kubenswrapper[4948]: I0220 08:07:40.721432 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:40 crc kubenswrapper[4948]: I0220 08:07:40.721529 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:40 crc kubenswrapper[4948]: I0220 08:07:40.721456 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:40 crc kubenswrapper[4948]: E0220 08:07:40.721783 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:40 crc kubenswrapper[4948]: E0220 08:07:40.721893 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:40 crc kubenswrapper[4948]: E0220 08:07:40.721670 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:41 crc kubenswrapper[4948]: I0220 08:07:41.721957 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:41 crc kubenswrapper[4948]: E0220 08:07:41.725092 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:41 crc kubenswrapper[4948]: I0220 08:07:41.726893 4948 scope.go:117] "RemoveContainer" containerID="7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b" Feb 20 08:07:41 crc kubenswrapper[4948]: E0220 08:07:41.727371 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" Feb 20 08:07:42 crc kubenswrapper[4948]: I0220 08:07:42.722531 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:42 crc kubenswrapper[4948]: I0220 08:07:42.722540 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:42 crc kubenswrapper[4948]: E0220 08:07:42.722797 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:42 crc kubenswrapper[4948]: E0220 08:07:42.722934 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:42 crc kubenswrapper[4948]: I0220 08:07:42.722540 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:42 crc kubenswrapper[4948]: E0220 08:07:42.724173 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:43 crc kubenswrapper[4948]: I0220 08:07:43.722117 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:43 crc kubenswrapper[4948]: E0220 08:07:43.722578 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:44 crc kubenswrapper[4948]: I0220 08:07:44.721843 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:44 crc kubenswrapper[4948]: I0220 08:07:44.721850 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:44 crc kubenswrapper[4948]: I0220 08:07:44.721915 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:44 crc kubenswrapper[4948]: E0220 08:07:44.722554 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:44 crc kubenswrapper[4948]: E0220 08:07:44.722759 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:44 crc kubenswrapper[4948]: E0220 08:07:44.722935 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:45 crc kubenswrapper[4948]: I0220 08:07:45.721683 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:45 crc kubenswrapper[4948]: E0220 08:07:45.721894 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:46 crc kubenswrapper[4948]: I0220 08:07:46.722327 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:46 crc kubenswrapper[4948]: I0220 08:07:46.722401 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:46 crc kubenswrapper[4948]: I0220 08:07:46.722328 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:46 crc kubenswrapper[4948]: E0220 08:07:46.722546 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:46 crc kubenswrapper[4948]: E0220 08:07:46.722660 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:46 crc kubenswrapper[4948]: E0220 08:07:46.722759 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:47 crc kubenswrapper[4948]: I0220 08:07:47.721854 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:47 crc kubenswrapper[4948]: E0220 08:07:47.722437 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:48 crc kubenswrapper[4948]: I0220 08:07:48.473572 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x7s56_b96124e4-0a74-4578-9142-fd728eb9f99e/kube-multus/1.log" Feb 20 08:07:48 crc kubenswrapper[4948]: I0220 08:07:48.475249 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x7s56_b96124e4-0a74-4578-9142-fd728eb9f99e/kube-multus/0.log" Feb 20 08:07:48 crc kubenswrapper[4948]: I0220 08:07:48.475495 4948 generic.go:334] "Generic (PLEG): container finished" podID="b96124e4-0a74-4578-9142-fd728eb9f99e" containerID="2f4093ec9d71181c58ad80a8c4ee3692148590df14aadaec96c0f54a843d1f8c" exitCode=1 Feb 20 08:07:48 crc kubenswrapper[4948]: I0220 08:07:48.475605 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x7s56" event={"ID":"b96124e4-0a74-4578-9142-fd728eb9f99e","Type":"ContainerDied","Data":"2f4093ec9d71181c58ad80a8c4ee3692148590df14aadaec96c0f54a843d1f8c"} Feb 20 08:07:48 crc kubenswrapper[4948]: I0220 08:07:48.475892 4948 scope.go:117] "RemoveContainer" containerID="34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e" Feb 20 08:07:48 crc kubenswrapper[4948]: I0220 08:07:48.476516 4948 scope.go:117] "RemoveContainer" containerID="2f4093ec9d71181c58ad80a8c4ee3692148590df14aadaec96c0f54a843d1f8c" Feb 20 08:07:48 crc kubenswrapper[4948]: E0220 08:07:48.476797 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-x7s56_openshift-multus(b96124e4-0a74-4578-9142-fd728eb9f99e)\"" pod="openshift-multus/multus-x7s56" podUID="b96124e4-0a74-4578-9142-fd728eb9f99e" Feb 20 08:07:48 crc kubenswrapper[4948]: I0220 08:07:48.721533 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:48 crc kubenswrapper[4948]: I0220 08:07:48.721557 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:48 crc kubenswrapper[4948]: E0220 08:07:48.721740 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:48 crc kubenswrapper[4948]: I0220 08:07:48.721563 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:48 crc kubenswrapper[4948]: E0220 08:07:48.721951 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:48 crc kubenswrapper[4948]: E0220 08:07:48.722192 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:49 crc kubenswrapper[4948]: I0220 08:07:49.482542 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x7s56_b96124e4-0a74-4578-9142-fd728eb9f99e/kube-multus/1.log" Feb 20 08:07:49 crc kubenswrapper[4948]: I0220 08:07:49.721794 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:49 crc kubenswrapper[4948]: E0220 08:07:49.722053 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:50 crc kubenswrapper[4948]: I0220 08:07:50.722158 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:50 crc kubenswrapper[4948]: I0220 08:07:50.722280 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:50 crc kubenswrapper[4948]: E0220 08:07:50.722380 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:50 crc kubenswrapper[4948]: I0220 08:07:50.722469 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:50 crc kubenswrapper[4948]: E0220 08:07:50.723299 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:50 crc kubenswrapper[4948]: E0220 08:07:50.723376 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:51 crc kubenswrapper[4948]: E0220 08:07:51.696518 4948 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Feb 20 08:07:51 crc kubenswrapper[4948]: I0220 08:07:51.722020 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:51 crc kubenswrapper[4948]: E0220 08:07:51.723332 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:52 crc kubenswrapper[4948]: E0220 08:07:52.295668 4948 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 20 08:07:52 crc kubenswrapper[4948]: I0220 08:07:52.721469 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:52 crc kubenswrapper[4948]: I0220 08:07:52.721493 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:52 crc kubenswrapper[4948]: E0220 08:07:52.721708 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:52 crc kubenswrapper[4948]: E0220 08:07:52.721799 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:52 crc kubenswrapper[4948]: I0220 08:07:52.721507 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:52 crc kubenswrapper[4948]: E0220 08:07:52.722560 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:53 crc kubenswrapper[4948]: I0220 08:07:53.722695 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:53 crc kubenswrapper[4948]: E0220 08:07:53.723004 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:53 crc kubenswrapper[4948]: I0220 08:07:53.723108 4948 scope.go:117] "RemoveContainer" containerID="7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b" Feb 20 08:07:53 crc kubenswrapper[4948]: E0220 08:07:53.723513 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" Feb 20 08:07:54 crc kubenswrapper[4948]: I0220 08:07:54.722386 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:54 crc kubenswrapper[4948]: I0220 08:07:54.722436 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:54 crc kubenswrapper[4948]: I0220 08:07:54.722397 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:54 crc kubenswrapper[4948]: E0220 08:07:54.722540 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:54 crc kubenswrapper[4948]: E0220 08:07:54.722782 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:54 crc kubenswrapper[4948]: E0220 08:07:54.722880 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:55 crc kubenswrapper[4948]: I0220 08:07:55.722372 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:55 crc kubenswrapper[4948]: E0220 08:07:55.722596 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:56 crc kubenswrapper[4948]: I0220 08:07:56.722205 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:56 crc kubenswrapper[4948]: I0220 08:07:56.722232 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:56 crc kubenswrapper[4948]: E0220 08:07:56.722354 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:56 crc kubenswrapper[4948]: I0220 08:07:56.722400 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:56 crc kubenswrapper[4948]: E0220 08:07:56.722578 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:56 crc kubenswrapper[4948]: E0220 08:07:56.722678 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:57 crc kubenswrapper[4948]: E0220 08:07:57.296957 4948 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 20 08:07:57 crc kubenswrapper[4948]: I0220 08:07:57.722519 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:57 crc kubenswrapper[4948]: E0220 08:07:57.722776 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:58 crc kubenswrapper[4948]: I0220 08:07:58.721864 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:58 crc kubenswrapper[4948]: I0220 08:07:58.722044 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:58 crc kubenswrapper[4948]: I0220 08:07:58.722396 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:58 crc kubenswrapper[4948]: E0220 08:07:58.722368 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:58 crc kubenswrapper[4948]: E0220 08:07:58.722679 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:58 crc kubenswrapper[4948]: E0220 08:07:58.722802 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:58 crc kubenswrapper[4948]: I0220 08:07:58.723336 4948 scope.go:117] "RemoveContainer" containerID="2f4093ec9d71181c58ad80a8c4ee3692148590df14aadaec96c0f54a843d1f8c" Feb 20 08:07:59 crc kubenswrapper[4948]: I0220 08:07:59.524424 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x7s56_b96124e4-0a74-4578-9142-fd728eb9f99e/kube-multus/1.log" Feb 20 08:07:59 crc kubenswrapper[4948]: I0220 08:07:59.524788 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x7s56" event={"ID":"b96124e4-0a74-4578-9142-fd728eb9f99e","Type":"ContainerStarted","Data":"4ecf8c7c9ed34410ff4619b18254e2eb90170911300e6cd6a7284e84e8937246"} Feb 20 08:07:59 crc kubenswrapper[4948]: I0220 08:07:59.722292 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:59 crc kubenswrapper[4948]: E0220 08:07:59.722510 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:08:00 crc kubenswrapper[4948]: I0220 08:08:00.721705 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:08:00 crc kubenswrapper[4948]: I0220 08:08:00.721812 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:08:00 crc kubenswrapper[4948]: I0220 08:08:00.721946 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:08:00 crc kubenswrapper[4948]: E0220 08:08:00.722292 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:08:00 crc kubenswrapper[4948]: E0220 08:08:00.722749 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:08:00 crc kubenswrapper[4948]: E0220 08:08:00.722165 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:08:01 crc kubenswrapper[4948]: I0220 08:08:01.721707 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:08:01 crc kubenswrapper[4948]: E0220 08:08:01.723543 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:08:02 crc kubenswrapper[4948]: E0220 08:08:02.298519 4948 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 20 08:08:02 crc kubenswrapper[4948]: I0220 08:08:02.722218 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:08:02 crc kubenswrapper[4948]: I0220 08:08:02.722258 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:08:02 crc kubenswrapper[4948]: I0220 08:08:02.722241 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:08:02 crc kubenswrapper[4948]: E0220 08:08:02.722433 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:08:02 crc kubenswrapper[4948]: E0220 08:08:02.722595 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:08:02 crc kubenswrapper[4948]: E0220 08:08:02.722731 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:08:03 crc kubenswrapper[4948]: I0220 08:08:03.722445 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:08:03 crc kubenswrapper[4948]: E0220 08:08:03.722673 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:08:04 crc kubenswrapper[4948]: I0220 08:08:04.722013 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:08:04 crc kubenswrapper[4948]: I0220 08:08:04.722076 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:08:04 crc kubenswrapper[4948]: I0220 08:08:04.722262 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:08:04 crc kubenswrapper[4948]: E0220 08:08:04.722560 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:08:04 crc kubenswrapper[4948]: E0220 08:08:04.722447 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:08:04 crc kubenswrapper[4948]: E0220 08:08:04.722246 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:08:05 crc kubenswrapper[4948]: I0220 08:08:05.721923 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:08:05 crc kubenswrapper[4948]: E0220 08:08:05.722213 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:08:06 crc kubenswrapper[4948]: I0220 08:08:06.721912 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:08:06 crc kubenswrapper[4948]: I0220 08:08:06.722031 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:08:06 crc kubenswrapper[4948]: I0220 08:08:06.721915 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:08:06 crc kubenswrapper[4948]: E0220 08:08:06.722187 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:08:06 crc kubenswrapper[4948]: E0220 08:08:06.722403 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:08:06 crc kubenswrapper[4948]: E0220 08:08:06.722558 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:08:06 crc kubenswrapper[4948]: I0220 08:08:06.723648 4948 scope.go:117] "RemoveContainer" containerID="7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b" Feb 20 08:08:07 crc kubenswrapper[4948]: E0220 08:08:07.300858 4948 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 20 08:08:07 crc kubenswrapper[4948]: I0220 08:08:07.557150 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovnkube-controller/3.log" Feb 20 08:08:07 crc kubenswrapper[4948]: I0220 08:08:07.560418 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerStarted","Data":"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a"} Feb 20 08:08:07 crc kubenswrapper[4948]: I0220 08:08:07.560880 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:08:07 crc kubenswrapper[4948]: I0220 08:08:07.602489 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podStartSLOduration=113.602455301 podStartE2EDuration="1m53.602455301s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:07.600146854 +0000 UTC m=+136.574641724" watchObservedRunningTime="2026-02-20 08:08:07.602455301 +0000 UTC m=+136.576950161" Feb 20 08:08:07 crc kubenswrapper[4948]: I0220 08:08:07.721941 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:08:07 crc kubenswrapper[4948]: E0220 08:08:07.722178 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:08:07 crc kubenswrapper[4948]: I0220 08:08:07.781854 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-xd86g"] Feb 20 08:08:07 crc kubenswrapper[4948]: I0220 08:08:07.782056 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:08:07 crc kubenswrapper[4948]: E0220 08:08:07.782254 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:08:08 crc kubenswrapper[4948]: I0220 08:08:08.722318 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:08:08 crc kubenswrapper[4948]: I0220 08:08:08.722318 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:08:08 crc kubenswrapper[4948]: E0220 08:08:08.722567 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:08:08 crc kubenswrapper[4948]: E0220 08:08:08.722741 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:08:09 crc kubenswrapper[4948]: I0220 08:08:09.722064 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:08:09 crc kubenswrapper[4948]: I0220 08:08:09.722170 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:08:09 crc kubenswrapper[4948]: E0220 08:08:09.722263 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:08:09 crc kubenswrapper[4948]: E0220 08:08:09.722420 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:08:10 crc kubenswrapper[4948]: I0220 08:08:10.722517 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:08:10 crc kubenswrapper[4948]: I0220 08:08:10.722568 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:08:10 crc kubenswrapper[4948]: E0220 08:08:10.722742 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:08:10 crc kubenswrapper[4948]: E0220 08:08:10.722884 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:08:11 crc kubenswrapper[4948]: I0220 08:08:11.722203 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:08:11 crc kubenswrapper[4948]: I0220 08:08:11.722309 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:08:11 crc kubenswrapper[4948]: E0220 08:08:11.724027 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:08:11 crc kubenswrapper[4948]: E0220 08:08:11.724249 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:08:12 crc kubenswrapper[4948]: I0220 08:08:12.722394 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:08:12 crc kubenswrapper[4948]: I0220 08:08:12.722522 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:08:12 crc kubenswrapper[4948]: I0220 08:08:12.725772 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 20 08:08:12 crc kubenswrapper[4948]: I0220 08:08:12.726786 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 20 08:08:12 crc kubenswrapper[4948]: I0220 08:08:12.726797 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 20 08:08:12 crc kubenswrapper[4948]: I0220 08:08:12.727030 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 20 08:08:13 crc kubenswrapper[4948]: I0220 08:08:13.721474 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:08:13 crc kubenswrapper[4948]: I0220 08:08:13.721565 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:08:13 crc kubenswrapper[4948]: I0220 08:08:13.726231 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 20 08:08:13 crc kubenswrapper[4948]: I0220 08:08:13.726335 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 20 08:08:20 crc kubenswrapper[4948]: I0220 08:08:20.592581 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:08:20 crc kubenswrapper[4948]: I0220 08:08:20.592641 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:08:20 crc kubenswrapper[4948]: I0220 08:08:20.594178 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:08:20 crc kubenswrapper[4948]: I0220 08:08:20.600845 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:08:20 crc kubenswrapper[4948]: I0220 08:08:20.693885 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:20 crc kubenswrapper[4948]: E0220 08:08:20.694182 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:10:22.694143079 +0000 UTC m=+271.668637949 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:20 crc kubenswrapper[4948]: I0220 08:08:20.694545 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:08:20 crc kubenswrapper[4948]: I0220 08:08:20.694629 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:08:20 crc kubenswrapper[4948]: I0220 08:08:20.700474 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:08:20 crc kubenswrapper[4948]: I0220 08:08:20.700575 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:08:20 crc kubenswrapper[4948]: I0220 08:08:20.847811 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:08:20 crc kubenswrapper[4948]: I0220 08:08:20.861315 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:08:20 crc kubenswrapper[4948]: I0220 08:08:20.955405 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:08:21 crc kubenswrapper[4948]: W0220 08:08:21.167423 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-44f5ba7b5021740fa36055309a0fed6bca4188708200cd5ea09e952a580b5810 WatchSource:0}: Error finding container 44f5ba7b5021740fa36055309a0fed6bca4188708200cd5ea09e952a580b5810: Status 404 returned error can't find the container with id 44f5ba7b5021740fa36055309a0fed6bca4188708200cd5ea09e952a580b5810 Feb 20 08:08:21 crc kubenswrapper[4948]: W0220 08:08:21.168384 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-88ea25004576fb4b079f000bc89eaff31a4353a7cea7a707f7fddd5a0870b507 WatchSource:0}: Error finding container 88ea25004576fb4b079f000bc89eaff31a4353a7cea7a707f7fddd5a0870b507: Status 404 returned error can't find the container with id 88ea25004576fb4b079f000bc89eaff31a4353a7cea7a707f7fddd5a0870b507 Feb 20 08:08:21 crc kubenswrapper[4948]: W0220 08:08:21.206477 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-ac2e7e878b2a17066ee69c643bf29d8a5bef205795d3458c0b19eb26a5d75d5f WatchSource:0}: Error finding container ac2e7e878b2a17066ee69c643bf29d8a5bef205795d3458c0b19eb26a5d75d5f: Status 404 returned error can't find the container with id ac2e7e878b2a17066ee69c643bf29d8a5bef205795d3458c0b19eb26a5d75d5f Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.301559 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.338281 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-jfst2"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.338778 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.343958 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.344003 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.344271 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.344380 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.346441 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.346705 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.346939 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.347062 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.352105 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.353709 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.354578 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.355607 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.356657 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.360872 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.362483 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.363007 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.363389 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.363943 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.364321 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7x85x"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.364633 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.370046 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.380732 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.381336 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.382477 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.382734 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-fhfhr"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.382955 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fl47j"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.383246 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fl47j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.383365 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7x85x" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.383548 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.384155 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-klsxg"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.384418 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.384666 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.384720 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-klsxg" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.388351 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.389245 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-fthdq"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.389617 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.389757 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.393180 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.393743 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.394113 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.394334 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.394474 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.394645 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.394768 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.395276 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.395445 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.397685 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.398233 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-qtlw7"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.398600 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.399744 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-xjbrz"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.399763 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.399846 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.400450 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.400579 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.400236 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.400384 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.402954 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c7c6bf9e-0846-4391-9126-a00d68b2e627-etcd-client\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.403008 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7c6bf9e-0846-4391-9126-a00d68b2e627-config\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.403037 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c7c6bf9e-0846-4391-9126-a00d68b2e627-encryption-config\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.403067 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c7c6bf9e-0846-4391-9126-a00d68b2e627-image-import-ca\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.403089 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c7c6bf9e-0846-4391-9126-a00d68b2e627-node-pullsecrets\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.403108 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c7c6bf9e-0846-4391-9126-a00d68b2e627-trusted-ca-bundle\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.403126 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hb9hj\" (UniqueName: \"kubernetes.io/projected/c7c6bf9e-0846-4391-9126-a00d68b2e627-kube-api-access-hb9hj\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.403149 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c7c6bf9e-0846-4391-9126-a00d68b2e627-audit\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.403167 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c7c6bf9e-0846-4391-9126-a00d68b2e627-etcd-serving-ca\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.403198 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7c6bf9e-0846-4391-9126-a00d68b2e627-serving-cert\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.403215 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c7c6bf9e-0846-4391-9126-a00d68b2e627-audit-dir\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.406100 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nzj8d"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.406506 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.406913 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vtkbz"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.407316 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vtkbz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.407839 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-xjbrz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.408161 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nzj8d" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.408466 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.409089 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-cc77j"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.409902 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-cc77j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.410295 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.410782 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.411351 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-44l7j"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.411833 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.412185 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.412420 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.412604 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.413181 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.413398 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.413530 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.414125 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.414375 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.414580 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.414664 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.414792 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-k8zcr"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.414123 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.414387 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.414897 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.415671 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.414923 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.414961 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.416458 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qfgtj"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.417137 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.417243 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-44l7j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.416608 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.416659 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.415048 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.415086 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.415082 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.414995 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.416907 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.416951 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.430098 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgm85"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.433131 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgm85" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.439812 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.440312 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.441315 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-x8fgj"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.441639 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.441641 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.442011 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.459557 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.463242 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.463452 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.463541 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.463610 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.463908 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.464710 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-x8fgj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.465114 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.465386 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.465467 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.465535 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.466664 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.467817 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.467957 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.469948 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.470074 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.470261 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.470478 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.470749 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.472195 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.472351 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.472386 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qgskm"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.472420 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.472575 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.472749 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.472900 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.472899 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-sc5pm"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.472996 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qgskm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.473099 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.473159 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.473286 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.473322 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.473516 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.474034 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.474108 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.474206 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.474359 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.474450 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.474517 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.478045 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-grnfc"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.478639 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-grnfc" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.479015 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.479355 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.481008 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.484750 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sjnxm"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.485569 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.489039 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.498551 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.498628 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.499706 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.500139 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-7wm8c"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.500939 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.501269 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-bcsmj"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.501883 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-bcsmj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.505262 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.506712 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.507006 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.507371 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.507934 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.508126 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.510722 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.511198 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.511846 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.512314 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.513001 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514178 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cf612779-d96d-48f1-9958-5f92ac6e3d40-serving-cert\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514201 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cf612779-d96d-48f1-9958-5f92ac6e3d40-etcd-client\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514219 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkhpt\" (UniqueName: \"kubernetes.io/projected/1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72-kube-api-access-gkhpt\") pod \"machine-config-controller-84d6567774-8rrhs\" (UID: \"1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514241 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-oauth-config\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514262 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e029f4ab-6a61-4fcf-9997-b51c1cbe5675-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-nzj8d\" (UID: \"e029f4ab-6a61-4fcf-9997-b51c1cbe5675\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nzj8d" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514276 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2c59c4e2-97ae-494a-b33a-a542266cf233-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514295 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-service-ca\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514309 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nq9dr\" (UniqueName: \"kubernetes.io/projected/a149fe37-c748-4120-9116-1da4b680d880-kube-api-access-nq9dr\") pod \"control-plane-machine-set-operator-78cbb6b69f-bgm85\" (UID: \"a149fe37-c748-4120-9116-1da4b680d880\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgm85" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514326 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e029f4ab-6a61-4fcf-9997-b51c1cbe5675-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-nzj8d\" (UID: \"e029f4ab-6a61-4fcf-9997-b51c1cbe5675\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nzj8d" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514342 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fabbac48-e5db-4057-b9a5-69118c38c667-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-vtkbz\" (UID: \"fabbac48-e5db-4057-b9a5-69118c38c667\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vtkbz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514357 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/5bc3d0cd-eacd-4b25-9acd-853e49db7b47-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-fthdq\" (UID: \"5bc3d0cd-eacd-4b25-9acd-853e49db7b47\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514373 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mz9sw\" (UniqueName: \"kubernetes.io/projected/de9b95d8-fa65-438c-bef9-2aba46044ec1-kube-api-access-mz9sw\") pod \"cluster-samples-operator-665b6dd947-7x85x\" (UID: \"de9b95d8-fa65-438c-bef9-2aba46044ec1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7x85x" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514388 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b1099221-b48f-4756-b133-0fed16b1e225-images\") pod \"machine-config-operator-74547568cd-lrdf2\" (UID: \"b1099221-b48f-4756-b133-0fed16b1e225\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514402 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsbfs\" (UniqueName: \"kubernetes.io/projected/21db7edd-6db3-4afa-b470-0016a9c5afb9-kube-api-access-vsbfs\") pod \"controller-manager-879f6c89f-fhfhr\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514418 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b1099221-b48f-4756-b133-0fed16b1e225-auth-proxy-config\") pod \"machine-config-operator-74547568cd-lrdf2\" (UID: \"b1099221-b48f-4756-b133-0fed16b1e225\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514436 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21db7edd-6db3-4afa-b470-0016a9c5afb9-serving-cert\") pod \"controller-manager-879f6c89f-fhfhr\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514453 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02036666-815d-4282-86f0-1b4163ea7cf9-config\") pod \"machine-approver-56656f9798-v4kfs\" (UID: \"02036666-815d-4282-86f0-1b4163ea7cf9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514469 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c38c7edf-ca04-4ea4-b15f-d83dea3f545e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-fl47j\" (UID: \"c38c7edf-ca04-4ea4-b15f-d83dea3f545e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fl47j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514486 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-oauth-serving-cert\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514502 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkj4v\" (UniqueName: \"kubernetes.io/projected/05fe2d3b-0a53-4bbe-b0aa-be07acb159bb-kube-api-access-jkj4v\") pod \"authentication-operator-69f744f599-qtlw7\" (UID: \"05fe2d3b-0a53-4bbe-b0aa-be07acb159bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514520 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/02036666-815d-4282-86f0-1b4163ea7cf9-machine-approver-tls\") pod \"machine-approver-56656f9798-v4kfs\" (UID: \"02036666-815d-4282-86f0-1b4163ea7cf9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514536 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bc3d0cd-eacd-4b25-9acd-853e49db7b47-config\") pod \"machine-api-operator-5694c8668f-fthdq\" (UID: \"5bc3d0cd-eacd-4b25-9acd-853e49db7b47\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514560 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-serving-cert\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514584 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/01d95eeb-e421-4fe2-a24a-44c3014f1f6e-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-thcgl\" (UID: \"01d95eeb-e421-4fe2-a24a-44c3014f1f6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514603 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c59c4e2-97ae-494a-b33a-a542266cf233-serving-cert\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514624 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgccc\" (UniqueName: \"kubernetes.io/projected/b1099221-b48f-4756-b133-0fed16b1e225-kube-api-access-lgccc\") pod \"machine-config-operator-74547568cd-lrdf2\" (UID: \"b1099221-b48f-4756-b133-0fed16b1e225\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514642 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05fe2d3b-0a53-4bbe-b0aa-be07acb159bb-config\") pod \"authentication-operator-69f744f599-qtlw7\" (UID: \"05fe2d3b-0a53-4bbe-b0aa-be07acb159bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514663 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c59c4e2-97ae-494a-b33a-a542266cf233-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514682 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mr8r\" (UniqueName: \"kubernetes.io/projected/5bc3d0cd-eacd-4b25-9acd-853e49db7b47-kube-api-access-5mr8r\") pod \"machine-api-operator-5694c8668f-fthdq\" (UID: \"5bc3d0cd-eacd-4b25-9acd-853e49db7b47\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514703 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6c070b67-c2ca-459b-a1b7-813a8833e27e-client-ca\") pod \"route-controller-manager-6576b87f9c-qljmk\" (UID: \"6c070b67-c2ca-459b-a1b7-813a8833e27e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514725 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/de9b95d8-fa65-438c-bef9-2aba46044ec1-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-7x85x\" (UID: \"de9b95d8-fa65-438c-bef9-2aba46044ec1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7x85x" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514745 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-client-ca\") pod \"controller-manager-879f6c89f-fhfhr\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514761 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8rrhs\" (UID: \"1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514776 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-config\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514791 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf612779-d96d-48f1-9958-5f92ac6e3d40-config\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514806 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bt5g\" (UniqueName: \"kubernetes.io/projected/cf612779-d96d-48f1-9958-5f92ac6e3d40-kube-api-access-2bt5g\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514826 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2c59c4e2-97ae-494a-b33a-a542266cf233-audit-policies\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514843 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9x9v\" (UniqueName: \"kubernetes.io/projected/2c59c4e2-97ae-494a-b33a-a542266cf233-kube-api-access-c9x9v\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514860 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/01d95eeb-e421-4fe2-a24a-44c3014f1f6e-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-thcgl\" (UID: \"01d95eeb-e421-4fe2-a24a-44c3014f1f6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514877 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94nxl\" (UniqueName: \"kubernetes.io/projected/e029f4ab-6a61-4fcf-9997-b51c1cbe5675-kube-api-access-94nxl\") pod \"kube-storage-version-migrator-operator-b67b599dd-nzj8d\" (UID: \"e029f4ab-6a61-4fcf-9997-b51c1cbe5675\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nzj8d" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514893 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/0b98dbb3-7986-475d-8028-0879d2fed2af-available-featuregates\") pod \"openshift-config-operator-7777fb866f-44l7j\" (UID: \"0b98dbb3-7986-475d-8028-0879d2fed2af\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-44l7j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514911 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhx6j\" (UniqueName: \"kubernetes.io/projected/bbc3437e-8729-46f8-aee2-ad3a4679097f-kube-api-access-lhx6j\") pod \"console-operator-58897d9998-xjbrz\" (UID: \"bbc3437e-8729-46f8-aee2-ad3a4679097f\") " pod="openshift-console-operator/console-operator-58897d9998-xjbrz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514934 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7c6bf9e-0846-4391-9126-a00d68b2e627-serving-cert\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514951 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c7c6bf9e-0846-4391-9126-a00d68b2e627-audit-dir\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514984 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbc3437e-8729-46f8-aee2-ad3a4679097f-config\") pod \"console-operator-58897d9998-xjbrz\" (UID: \"bbc3437e-8729-46f8-aee2-ad3a4679097f\") " pod="openshift-console-operator/console-operator-58897d9998-xjbrz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515001 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s94wm\" (UniqueName: \"kubernetes.io/projected/02036666-815d-4282-86f0-1b4163ea7cf9-kube-api-access-s94wm\") pod \"machine-approver-56656f9798-v4kfs\" (UID: \"02036666-815d-4282-86f0-1b4163ea7cf9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515017 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-config\") pod \"controller-manager-879f6c89f-fhfhr\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515033 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2c59c4e2-97ae-494a-b33a-a542266cf233-etcd-client\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515054 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2c59c4e2-97ae-494a-b33a-a542266cf233-encryption-config\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515069 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c38c7edf-ca04-4ea4-b15f-d83dea3f545e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-fl47j\" (UID: \"c38c7edf-ca04-4ea4-b15f-d83dea3f545e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fl47j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515084 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c7c6bf9e-0846-4391-9126-a00d68b2e627-etcd-client\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515100 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7c6bf9e-0846-4391-9126-a00d68b2e627-config\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515118 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk7rx\" (UniqueName: \"kubernetes.io/projected/0b81d26a-f92a-4602-89ef-5dd9fb24a32f-kube-api-access-lk7rx\") pod \"openshift-controller-manager-operator-756b6f6bc6-klsxg\" (UID: \"0b81d26a-f92a-4602-89ef-5dd9fb24a32f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-klsxg" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515133 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fabbac48-e5db-4057-b9a5-69118c38c667-config\") pod \"kube-apiserver-operator-766d6c64bb-vtkbz\" (UID: \"fabbac48-e5db-4057-b9a5-69118c38c667\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vtkbz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515147 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2c59c4e2-97ae-494a-b33a-a542266cf233-audit-dir\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515161 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5bc3d0cd-eacd-4b25-9acd-853e49db7b47-images\") pod \"machine-api-operator-5694c8668f-fthdq\" (UID: \"5bc3d0cd-eacd-4b25-9acd-853e49db7b47\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515175 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05fe2d3b-0a53-4bbe-b0aa-be07acb159bb-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-qtlw7\" (UID: \"05fe2d3b-0a53-4bbe-b0aa-be07acb159bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515190 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05fe2d3b-0a53-4bbe-b0aa-be07acb159bb-serving-cert\") pod \"authentication-operator-69f744f599-qtlw7\" (UID: \"05fe2d3b-0a53-4bbe-b0aa-be07acb159bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515205 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b1099221-b48f-4756-b133-0fed16b1e225-proxy-tls\") pod \"machine-config-operator-74547568cd-lrdf2\" (UID: \"b1099221-b48f-4756-b133-0fed16b1e225\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515223 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/cf612779-d96d-48f1-9958-5f92ac6e3d40-etcd-ca\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515237 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/cf612779-d96d-48f1-9958-5f92ac6e3d40-etcd-service-ca\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515252 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bbc3437e-8729-46f8-aee2-ad3a4679097f-trusted-ca\") pod \"console-operator-58897d9998-xjbrz\" (UID: \"bbc3437e-8729-46f8-aee2-ad3a4679097f\") " pod="openshift-console-operator/console-operator-58897d9998-xjbrz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515267 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c070b67-c2ca-459b-a1b7-813a8833e27e-config\") pod \"route-controller-manager-6576b87f9c-qljmk\" (UID: \"6c070b67-c2ca-459b-a1b7-813a8833e27e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515446 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c7c6bf9e-0846-4391-9126-a00d68b2e627-encryption-config\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515495 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-trusted-ca-bundle\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515513 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a149fe37-c748-4120-9116-1da4b680d880-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-bgm85\" (UID: \"a149fe37-c748-4120-9116-1da4b680d880\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgm85" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515531 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2v824\" (UniqueName: \"kubernetes.io/projected/01d95eeb-e421-4fe2-a24a-44c3014f1f6e-kube-api-access-2v824\") pod \"cluster-image-registry-operator-dc59b4c8b-thcgl\" (UID: \"01d95eeb-e421-4fe2-a24a-44c3014f1f6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515555 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fabbac48-e5db-4057-b9a5-69118c38c667-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-vtkbz\" (UID: \"fabbac48-e5db-4057-b9a5-69118c38c667\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vtkbz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515576 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c070b67-c2ca-459b-a1b7-813a8833e27e-serving-cert\") pod \"route-controller-manager-6576b87f9c-qljmk\" (UID: \"6c070b67-c2ca-459b-a1b7-813a8833e27e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515603 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c7c6bf9e-0846-4391-9126-a00d68b2e627-image-import-ca\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515627 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-fhfhr\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515648 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25nlb\" (UniqueName: \"kubernetes.io/projected/6c070b67-c2ca-459b-a1b7-813a8833e27e-kube-api-access-25nlb\") pod \"route-controller-manager-6576b87f9c-qljmk\" (UID: \"6c070b67-c2ca-459b-a1b7-813a8833e27e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515666 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9k4m\" (UniqueName: \"kubernetes.io/projected/c38c7edf-ca04-4ea4-b15f-d83dea3f545e-kube-api-access-m9k4m\") pod \"openshift-apiserver-operator-796bbdcf4f-fl47j\" (UID: \"c38c7edf-ca04-4ea4-b15f-d83dea3f545e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fl47j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515684 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c7c6bf9e-0846-4391-9126-a00d68b2e627-node-pullsecrets\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515700 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b98dbb3-7986-475d-8028-0879d2fed2af-serving-cert\") pod \"openshift-config-operator-7777fb866f-44l7j\" (UID: \"0b98dbb3-7986-475d-8028-0879d2fed2af\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-44l7j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515717 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c7c6bf9e-0846-4391-9126-a00d68b2e627-trusted-ca-bundle\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515735 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hb9hj\" (UniqueName: \"kubernetes.io/projected/c7c6bf9e-0846-4391-9126-a00d68b2e627-kube-api-access-hb9hj\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515752 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b81d26a-f92a-4602-89ef-5dd9fb24a32f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-klsxg\" (UID: \"0b81d26a-f92a-4602-89ef-5dd9fb24a32f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-klsxg" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515772 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccbnv\" (UniqueName: \"kubernetes.io/projected/0b98dbb3-7986-475d-8028-0879d2fed2af-kube-api-access-ccbnv\") pod \"openshift-config-operator-7777fb866f-44l7j\" (UID: \"0b98dbb3-7986-475d-8028-0879d2fed2af\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-44l7j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515791 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/01d95eeb-e421-4fe2-a24a-44c3014f1f6e-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-thcgl\" (UID: \"01d95eeb-e421-4fe2-a24a-44c3014f1f6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515807 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72-proxy-tls\") pod \"machine-config-controller-84d6567774-8rrhs\" (UID: \"1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515825 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xszfj\" (UniqueName: \"kubernetes.io/projected/3f4bc9ec-3eda-4d56-9696-df15c14d2e4a-kube-api-access-xszfj\") pod \"migrator-59844c95c7-cc77j\" (UID: \"3f4bc9ec-3eda-4d56-9696-df15c14d2e4a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-cc77j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515842 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7b5rg\" (UniqueName: \"kubernetes.io/projected/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-kube-api-access-7b5rg\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515859 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c7c6bf9e-0846-4391-9126-a00d68b2e627-audit\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515910 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/02036666-815d-4282-86f0-1b4163ea7cf9-auth-proxy-config\") pod \"machine-approver-56656f9798-v4kfs\" (UID: \"02036666-815d-4282-86f0-1b4163ea7cf9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515927 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c7c6bf9e-0846-4391-9126-a00d68b2e627-etcd-serving-ca\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515944 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b81d26a-f92a-4602-89ef-5dd9fb24a32f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-klsxg\" (UID: \"0b81d26a-f92a-4602-89ef-5dd9fb24a32f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-klsxg" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515959 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05fe2d3b-0a53-4bbe-b0aa-be07acb159bb-service-ca-bundle\") pod \"authentication-operator-69f744f599-qtlw7\" (UID: \"05fe2d3b-0a53-4bbe-b0aa-be07acb159bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515990 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bbc3437e-8729-46f8-aee2-ad3a4679097f-serving-cert\") pod \"console-operator-58897d9998-xjbrz\" (UID: \"bbc3437e-8729-46f8-aee2-ad3a4679097f\") " pod="openshift-console-operator/console-operator-58897d9998-xjbrz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.516860 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-tdhsc"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.517330 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5kfk4"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.517639 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-46sg8"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.521453 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c7c6bf9e-0846-4391-9126-a00d68b2e627-audit-dir\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.523236 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-46sg8" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.523845 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-tdhsc" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.524059 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5kfk4" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.524819 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7c6bf9e-0846-4391-9126-a00d68b2e627-serving-cert\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.530583 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c7c6bf9e-0846-4391-9126-a00d68b2e627-etcd-client\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.531141 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c7c6bf9e-0846-4391-9126-a00d68b2e627-image-import-ca\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.531254 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c7c6bf9e-0846-4391-9126-a00d68b2e627-node-pullsecrets\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.532341 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7c6bf9e-0846-4391-9126-a00d68b2e627-config\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.536297 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c7c6bf9e-0846-4391-9126-a00d68b2e627-etcd-serving-ca\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.536738 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c7c6bf9e-0846-4391-9126-a00d68b2e627-audit\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.538525 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c7c6bf9e-0846-4391-9126-a00d68b2e627-trusted-ca-bundle\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.538579 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rtnmx"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.542458 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.542473 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.544872 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c7c6bf9e-0846-4391-9126-a00d68b2e627-encryption-config\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.546737 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sjv8t"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.547002 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-rtnmx" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.548491 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-jfst2"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.548605 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.549924 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.560034 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-fhfhr"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.562751 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.564231 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.565268 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-klsxg"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.566753 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fl47j"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.569012 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-fthdq"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.570506 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.572426 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-xjbrz"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.574720 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7x85x"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.576321 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.577827 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-9ls2l"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.578749 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-9ls2l" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.579922 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-bcsmj"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.580842 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-44l7j"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.582292 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-bp2vx"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.583557 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.584036 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgm85"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.584045 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-bp2vx" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.585369 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-grnfc"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.586996 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-k8zcr"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.589660 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vtkbz"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.591636 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-qtlw7"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.593213 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nzj8d"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.594485 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.595757 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.600842 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-7wm8c"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.602048 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.603883 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qfgtj"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.605310 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.606678 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.608187 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-x8fgj"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.609349 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.610925 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.612286 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qgskm"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.613949 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sjnxm"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.615374 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-trh7g"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.616733 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bbc3437e-8729-46f8-aee2-ad3a4679097f-serving-cert\") pod \"console-operator-58897d9998-xjbrz\" (UID: \"bbc3437e-8729-46f8-aee2-ad3a4679097f\") " pod="openshift-console-operator/console-operator-58897d9998-xjbrz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.616775 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.616801 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e9f46a23-4535-4d64-ae76-065435f8f762-metrics-tls\") pod \"ingress-operator-5b745b69d9-6jlbw\" (UID: \"e9f46a23-4535-4d64-ae76-065435f8f762\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.616824 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b81d26a-f92a-4602-89ef-5dd9fb24a32f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-klsxg\" (UID: \"0b81d26a-f92a-4602-89ef-5dd9fb24a32f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-klsxg" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.616843 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05fe2d3b-0a53-4bbe-b0aa-be07acb159bb-service-ca-bundle\") pod \"authentication-operator-69f744f599-qtlw7\" (UID: \"05fe2d3b-0a53-4bbe-b0aa-be07acb159bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.616871 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e029f4ab-6a61-4fcf-9997-b51c1cbe5675-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-nzj8d\" (UID: \"e029f4ab-6a61-4fcf-9997-b51c1cbe5675\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nzj8d" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.616892 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cf612779-d96d-48f1-9958-5f92ac6e3d40-serving-cert\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.616915 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cf612779-d96d-48f1-9958-5f92ac6e3d40-etcd-client\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.616935 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkhpt\" (UniqueName: \"kubernetes.io/projected/1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72-kube-api-access-gkhpt\") pod \"machine-config-controller-84d6567774-8rrhs\" (UID: \"1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.616953 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-oauth-config\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.616982 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2c59c4e2-97ae-494a-b33a-a542266cf233-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617002 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c7b67dd6-4025-45ad-98f8-2e7f8ab4b573-srv-cert\") pod \"olm-operator-6b444d44fb-6zdkm\" (UID: \"c7b67dd6-4025-45ad-98f8-2e7f8ab4b573\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617036 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2l9z\" (UniqueName: \"kubernetes.io/projected/df73228e-df1f-4e9c-aafe-493bb743f98a-kube-api-access-q2l9z\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617062 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fabbac48-e5db-4057-b9a5-69118c38c667-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-vtkbz\" (UID: \"fabbac48-e5db-4057-b9a5-69118c38c667\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vtkbz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617087 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-service-ca\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617117 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nq9dr\" (UniqueName: \"kubernetes.io/projected/a149fe37-c748-4120-9116-1da4b680d880-kube-api-access-nq9dr\") pod \"control-plane-machine-set-operator-78cbb6b69f-bgm85\" (UID: \"a149fe37-c748-4120-9116-1da4b680d880\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgm85" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617143 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e029f4ab-6a61-4fcf-9997-b51c1cbe5675-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-nzj8d\" (UID: \"e029f4ab-6a61-4fcf-9997-b51c1cbe5675\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nzj8d" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617167 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b1099221-b48f-4756-b133-0fed16b1e225-images\") pod \"machine-config-operator-74547568cd-lrdf2\" (UID: \"b1099221-b48f-4756-b133-0fed16b1e225\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617184 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/5bc3d0cd-eacd-4b25-9acd-853e49db7b47-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-fthdq\" (UID: \"5bc3d0cd-eacd-4b25-9acd-853e49db7b47\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617202 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mz9sw\" (UniqueName: \"kubernetes.io/projected/de9b95d8-fa65-438c-bef9-2aba46044ec1-kube-api-access-mz9sw\") pod \"cluster-samples-operator-665b6dd947-7x85x\" (UID: \"de9b95d8-fa65-438c-bef9-2aba46044ec1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7x85x" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617228 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsbfs\" (UniqueName: \"kubernetes.io/projected/21db7edd-6db3-4afa-b470-0016a9c5afb9-kube-api-access-vsbfs\") pod \"controller-manager-879f6c89f-fhfhr\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617247 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617273 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02036666-815d-4282-86f0-1b4163ea7cf9-config\") pod \"machine-approver-56656f9798-v4kfs\" (UID: \"02036666-815d-4282-86f0-1b4163ea7cf9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617359 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b1099221-b48f-4756-b133-0fed16b1e225-auth-proxy-config\") pod \"machine-config-operator-74547568cd-lrdf2\" (UID: \"b1099221-b48f-4756-b133-0fed16b1e225\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617810 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21db7edd-6db3-4afa-b470-0016a9c5afb9-serving-cert\") pod \"controller-manager-879f6c89f-fhfhr\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617864 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmpf7\" (UniqueName: \"kubernetes.io/projected/e9f46a23-4535-4d64-ae76-065435f8f762-kube-api-access-lmpf7\") pod \"ingress-operator-5b745b69d9-6jlbw\" (UID: \"e9f46a23-4535-4d64-ae76-065435f8f762\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617904 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrvhw\" (UniqueName: \"kubernetes.io/projected/bc13d9fe-fd58-4d3b-9278-4f3e59da5976-kube-api-access-lrvhw\") pod \"downloads-7954f5f757-grnfc\" (UID: \"bc13d9fe-fd58-4d3b-9278-4f3e59da5976\") " pod="openshift-console/downloads-7954f5f757-grnfc" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617934 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c38c7edf-ca04-4ea4-b15f-d83dea3f545e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-fl47j\" (UID: \"c38c7edf-ca04-4ea4-b15f-d83dea3f545e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fl47j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617960 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-audit-policies\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618002 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/02036666-815d-4282-86f0-1b4163ea7cf9-machine-approver-tls\") pod \"machine-approver-56656f9798-v4kfs\" (UID: \"02036666-815d-4282-86f0-1b4163ea7cf9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618049 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-oauth-serving-cert\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618078 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkj4v\" (UniqueName: \"kubernetes.io/projected/05fe2d3b-0a53-4bbe-b0aa-be07acb159bb-kube-api-access-jkj4v\") pod \"authentication-operator-69f744f599-qtlw7\" (UID: \"05fe2d3b-0a53-4bbe-b0aa-be07acb159bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618103 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618131 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618152 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bc3d0cd-eacd-4b25-9acd-853e49db7b47-config\") pod \"machine-api-operator-5694c8668f-fthdq\" (UID: \"5bc3d0cd-eacd-4b25-9acd-853e49db7b47\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618173 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-serving-cert\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618192 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/01d95eeb-e421-4fe2-a24a-44c3014f1f6e-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-thcgl\" (UID: \"01d95eeb-e421-4fe2-a24a-44c3014f1f6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618208 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c59c4e2-97ae-494a-b33a-a542266cf233-serving-cert\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618226 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgccc\" (UniqueName: \"kubernetes.io/projected/b1099221-b48f-4756-b133-0fed16b1e225-kube-api-access-lgccc\") pod \"machine-config-operator-74547568cd-lrdf2\" (UID: \"b1099221-b48f-4756-b133-0fed16b1e225\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618244 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05fe2d3b-0a53-4bbe-b0aa-be07acb159bb-config\") pod \"authentication-operator-69f744f599-qtlw7\" (UID: \"05fe2d3b-0a53-4bbe-b0aa-be07acb159bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618262 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c59c4e2-97ae-494a-b33a-a542266cf233-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618282 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-client-ca\") pod \"controller-manager-879f6c89f-fhfhr\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618298 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mr8r\" (UniqueName: \"kubernetes.io/projected/5bc3d0cd-eacd-4b25-9acd-853e49db7b47-kube-api-access-5mr8r\") pod \"machine-api-operator-5694c8668f-fthdq\" (UID: \"5bc3d0cd-eacd-4b25-9acd-853e49db7b47\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618317 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6c070b67-c2ca-459b-a1b7-813a8833e27e-client-ca\") pod \"route-controller-manager-6576b87f9c-qljmk\" (UID: \"6c070b67-c2ca-459b-a1b7-813a8833e27e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618338 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hqgh\" (UniqueName: \"kubernetes.io/projected/6f1e0caa-03ee-4025-9f2a-6941090c178b-kube-api-access-6hqgh\") pod \"dns-operator-744455d44c-x8fgj\" (UID: \"6f1e0caa-03ee-4025-9f2a-6941090c178b\") " pod="openshift-dns-operator/dns-operator-744455d44c-x8fgj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618364 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/de9b95d8-fa65-438c-bef9-2aba46044ec1-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-7x85x\" (UID: \"de9b95d8-fa65-438c-bef9-2aba46044ec1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7x85x" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618383 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618400 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618418 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618438 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf612779-d96d-48f1-9958-5f92ac6e3d40-config\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618448 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02036666-815d-4282-86f0-1b4163ea7cf9-config\") pod \"machine-approver-56656f9798-v4kfs\" (UID: \"02036666-815d-4282-86f0-1b4163ea7cf9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618457 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8rrhs\" (UID: \"1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618500 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-config\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618520 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94nxl\" (UniqueName: \"kubernetes.io/projected/e029f4ab-6a61-4fcf-9997-b51c1cbe5675-kube-api-access-94nxl\") pod \"kube-storage-version-migrator-operator-b67b599dd-nzj8d\" (UID: \"e029f4ab-6a61-4fcf-9997-b51c1cbe5675\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nzj8d" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618538 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bt5g\" (UniqueName: \"kubernetes.io/projected/cf612779-d96d-48f1-9958-5f92ac6e3d40-kube-api-access-2bt5g\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618557 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2c59c4e2-97ae-494a-b33a-a542266cf233-audit-policies\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618564 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-gd69n"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618577 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9x9v\" (UniqueName: \"kubernetes.io/projected/2c59c4e2-97ae-494a-b33a-a542266cf233-kube-api-access-c9x9v\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618598 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2c59c4e2-97ae-494a-b33a-a542266cf233-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618821 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618568 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b81d26a-f92a-4602-89ef-5dd9fb24a32f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-klsxg\" (UID: \"0b81d26a-f92a-4602-89ef-5dd9fb24a32f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-klsxg" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618945 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/01d95eeb-e421-4fe2-a24a-44c3014f1f6e-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-thcgl\" (UID: \"01d95eeb-e421-4fe2-a24a-44c3014f1f6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618996 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/0b98dbb3-7986-475d-8028-0879d2fed2af-available-featuregates\") pod \"openshift-config-operator-7777fb866f-44l7j\" (UID: \"0b98dbb3-7986-475d-8028-0879d2fed2af\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-44l7j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619021 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e9f46a23-4535-4d64-ae76-065435f8f762-trusted-ca\") pod \"ingress-operator-5b745b69d9-6jlbw\" (UID: \"e9f46a23-4535-4d64-ae76-065435f8f762\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619042 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbc3437e-8729-46f8-aee2-ad3a4679097f-config\") pod \"console-operator-58897d9998-xjbrz\" (UID: \"bbc3437e-8729-46f8-aee2-ad3a4679097f\") " pod="openshift-console-operator/console-operator-58897d9998-xjbrz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619061 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhx6j\" (UniqueName: \"kubernetes.io/projected/bbc3437e-8729-46f8-aee2-ad3a4679097f-kube-api-access-lhx6j\") pod \"console-operator-58897d9998-xjbrz\" (UID: \"bbc3437e-8729-46f8-aee2-ad3a4679097f\") " pod="openshift-console-operator/console-operator-58897d9998-xjbrz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619079 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619098 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n92xj\" (UniqueName: \"kubernetes.io/projected/c7b67dd6-4025-45ad-98f8-2e7f8ab4b573-kube-api-access-n92xj\") pod \"olm-operator-6b444d44fb-6zdkm\" (UID: \"c7b67dd6-4025-45ad-98f8-2e7f8ab4b573\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619119 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s94wm\" (UniqueName: \"kubernetes.io/projected/02036666-815d-4282-86f0-1b4163ea7cf9-kube-api-access-s94wm\") pod \"machine-approver-56656f9798-v4kfs\" (UID: \"02036666-815d-4282-86f0-1b4163ea7cf9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619134 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-config\") pod \"controller-manager-879f6c89f-fhfhr\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619151 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2c59c4e2-97ae-494a-b33a-a542266cf233-etcd-client\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619168 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2c59c4e2-97ae-494a-b33a-a542266cf233-encryption-config\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619188 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c38c7edf-ca04-4ea4-b15f-d83dea3f545e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-fl47j\" (UID: \"c38c7edf-ca04-4ea4-b15f-d83dea3f545e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fl47j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619198 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b1099221-b48f-4756-b133-0fed16b1e225-images\") pod \"machine-config-operator-74547568cd-lrdf2\" (UID: \"b1099221-b48f-4756-b133-0fed16b1e225\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619214 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/df73228e-df1f-4e9c-aafe-493bb743f98a-audit-dir\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619247 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk7rx\" (UniqueName: \"kubernetes.io/projected/0b81d26a-f92a-4602-89ef-5dd9fb24a32f-kube-api-access-lk7rx\") pod \"openshift-controller-manager-operator-756b6f6bc6-klsxg\" (UID: \"0b81d26a-f92a-4602-89ef-5dd9fb24a32f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-klsxg" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619273 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fabbac48-e5db-4057-b9a5-69118c38c667-config\") pod \"kube-apiserver-operator-766d6c64bb-vtkbz\" (UID: \"fabbac48-e5db-4057-b9a5-69118c38c667\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vtkbz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619295 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2c59c4e2-97ae-494a-b33a-a542266cf233-audit-dir\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619311 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619307 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-service-ca\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619332 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b1099221-b48f-4756-b133-0fed16b1e225-proxy-tls\") pod \"machine-config-operator-74547568cd-lrdf2\" (UID: \"b1099221-b48f-4756-b133-0fed16b1e225\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619517 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5bc3d0cd-eacd-4b25-9acd-853e49db7b47-images\") pod \"machine-api-operator-5694c8668f-fthdq\" (UID: \"5bc3d0cd-eacd-4b25-9acd-853e49db7b47\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619538 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05fe2d3b-0a53-4bbe-b0aa-be07acb159bb-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-qtlw7\" (UID: \"05fe2d3b-0a53-4bbe-b0aa-be07acb159bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619556 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05fe2d3b-0a53-4bbe-b0aa-be07acb159bb-serving-cert\") pod \"authentication-operator-69f744f599-qtlw7\" (UID: \"05fe2d3b-0a53-4bbe-b0aa-be07acb159bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619582 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6f1e0caa-03ee-4025-9f2a-6941090c178b-metrics-tls\") pod \"dns-operator-744455d44c-x8fgj\" (UID: \"6f1e0caa-03ee-4025-9f2a-6941090c178b\") " pod="openshift-dns-operator/dns-operator-744455d44c-x8fgj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619598 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619619 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/cf612779-d96d-48f1-9958-5f92ac6e3d40-etcd-ca\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619636 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/cf612779-d96d-48f1-9958-5f92ac6e3d40-etcd-service-ca\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619652 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bbc3437e-8729-46f8-aee2-ad3a4679097f-trusted-ca\") pod \"console-operator-58897d9998-xjbrz\" (UID: \"bbc3437e-8729-46f8-aee2-ad3a4679097f\") " pod="openshift-console-operator/console-operator-58897d9998-xjbrz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619671 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a149fe37-c748-4120-9116-1da4b680d880-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-bgm85\" (UID: \"a149fe37-c748-4120-9116-1da4b680d880\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgm85" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619690 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c070b67-c2ca-459b-a1b7-813a8833e27e-config\") pod \"route-controller-manager-6576b87f9c-qljmk\" (UID: \"6c070b67-c2ca-459b-a1b7-813a8833e27e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619710 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-trusted-ca-bundle\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619733 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-fhfhr\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619754 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2v824\" (UniqueName: \"kubernetes.io/projected/01d95eeb-e421-4fe2-a24a-44c3014f1f6e-kube-api-access-2v824\") pod \"cluster-image-registry-operator-dc59b4c8b-thcgl\" (UID: \"01d95eeb-e421-4fe2-a24a-44c3014f1f6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619773 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fabbac48-e5db-4057-b9a5-69118c38c667-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-vtkbz\" (UID: \"fabbac48-e5db-4057-b9a5-69118c38c667\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vtkbz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619794 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c070b67-c2ca-459b-a1b7-813a8833e27e-serving-cert\") pod \"route-controller-manager-6576b87f9c-qljmk\" (UID: \"6c070b67-c2ca-459b-a1b7-813a8833e27e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619827 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25nlb\" (UniqueName: \"kubernetes.io/projected/6c070b67-c2ca-459b-a1b7-813a8833e27e-kube-api-access-25nlb\") pod \"route-controller-manager-6576b87f9c-qljmk\" (UID: \"6c070b67-c2ca-459b-a1b7-813a8833e27e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619825 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6c070b67-c2ca-459b-a1b7-813a8833e27e-client-ca\") pod \"route-controller-manager-6576b87f9c-qljmk\" (UID: \"6c070b67-c2ca-459b-a1b7-813a8833e27e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619846 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9k4m\" (UniqueName: \"kubernetes.io/projected/c38c7edf-ca04-4ea4-b15f-d83dea3f545e-kube-api-access-m9k4m\") pod \"openshift-apiserver-operator-796bbdcf4f-fl47j\" (UID: \"c38c7edf-ca04-4ea4-b15f-d83dea3f545e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fl47j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619868 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b98dbb3-7986-475d-8028-0879d2fed2af-serving-cert\") pod \"openshift-config-operator-7777fb866f-44l7j\" (UID: \"0b98dbb3-7986-475d-8028-0879d2fed2af\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-44l7j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619893 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b81d26a-f92a-4602-89ef-5dd9fb24a32f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-klsxg\" (UID: \"0b81d26a-f92a-4602-89ef-5dd9fb24a32f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-klsxg" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619915 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccbnv\" (UniqueName: \"kubernetes.io/projected/0b98dbb3-7986-475d-8028-0879d2fed2af-kube-api-access-ccbnv\") pod \"openshift-config-operator-7777fb866f-44l7j\" (UID: \"0b98dbb3-7986-475d-8028-0879d2fed2af\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-44l7j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619935 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619952 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c7b67dd6-4025-45ad-98f8-2e7f8ab4b573-profile-collector-cert\") pod \"olm-operator-6b444d44fb-6zdkm\" (UID: \"c7b67dd6-4025-45ad-98f8-2e7f8ab4b573\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.620011 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7b5rg\" (UniqueName: \"kubernetes.io/projected/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-kube-api-access-7b5rg\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.620032 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/01d95eeb-e421-4fe2-a24a-44c3014f1f6e-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-thcgl\" (UID: \"01d95eeb-e421-4fe2-a24a-44c3014f1f6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.620054 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72-proxy-tls\") pod \"machine-config-controller-84d6567774-8rrhs\" (UID: \"1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619692 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bc3d0cd-eacd-4b25-9acd-853e49db7b47-config\") pod \"machine-api-operator-5694c8668f-fthdq\" (UID: \"5bc3d0cd-eacd-4b25-9acd-853e49db7b47\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.621065 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e029f4ab-6a61-4fcf-9997-b51c1cbe5675-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-nzj8d\" (UID: \"e029f4ab-6a61-4fcf-9997-b51c1cbe5675\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nzj8d" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.621448 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-oauth-config\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.621861 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-config\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.621913 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xszfj\" (UniqueName: \"kubernetes.io/projected/3f4bc9ec-3eda-4d56-9696-df15c14d2e4a-kube-api-access-xszfj\") pod \"migrator-59844c95c7-cc77j\" (UID: \"3f4bc9ec-3eda-4d56-9696-df15c14d2e4a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-cc77j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.621942 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e9f46a23-4535-4d64-ae76-065435f8f762-bound-sa-token\") pod \"ingress-operator-5b745b69d9-6jlbw\" (UID: \"e9f46a23-4535-4d64-ae76-065435f8f762\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.621969 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/02036666-815d-4282-86f0-1b4163ea7cf9-auth-proxy-config\") pod \"machine-approver-56656f9798-v4kfs\" (UID: \"02036666-815d-4282-86f0-1b4163ea7cf9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.622025 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2c59c4e2-97ae-494a-b33a-a542266cf233-audit-policies\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.622483 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/02036666-815d-4282-86f0-1b4163ea7cf9-auth-proxy-config\") pod \"machine-approver-56656f9798-v4kfs\" (UID: \"02036666-815d-4282-86f0-1b4163ea7cf9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.622492 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cf612779-d96d-48f1-9958-5f92ac6e3d40-etcd-client\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.622768 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.622818 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8rrhs\" (UID: \"1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.622956 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e029f4ab-6a61-4fcf-9997-b51c1cbe5675-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-nzj8d\" (UID: \"e029f4ab-6a61-4fcf-9997-b51c1cbe5675\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nzj8d" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.623271 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-serving-cert\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.623673 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b1099221-b48f-4756-b133-0fed16b1e225-auth-proxy-config\") pod \"machine-config-operator-74547568cd-lrdf2\" (UID: \"b1099221-b48f-4756-b133-0fed16b1e225\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.623744 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/cf612779-d96d-48f1-9958-5f92ac6e3d40-etcd-ca\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.624214 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"6d33a120c6d6404bfdafb94ac3396204dc5d916bd0861e293668c90e52a5a928"} Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.624269 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5kfk4"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.624295 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rtnmx"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.624310 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"ac2e7e878b2a17066ee69c643bf29d8a5bef205795d3458c0b19eb26a5d75d5f"} Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.624324 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"448b4e1f87f68cbd93b97818fd8a7931f478e3d00887e8cf4aadad7e8d891215"} Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.624337 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"88ea25004576fb4b079f000bc89eaff31a4353a7cea7a707f7fddd5a0870b507"} Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.624622 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05fe2d3b-0a53-4bbe-b0aa-be07acb159bb-config\") pod \"authentication-operator-69f744f599-qtlw7\" (UID: \"05fe2d3b-0a53-4bbe-b0aa-be07acb159bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.624837 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bbc3437e-8729-46f8-aee2-ad3a4679097f-serving-cert\") pod \"console-operator-58897d9998-xjbrz\" (UID: \"bbc3437e-8729-46f8-aee2-ad3a4679097f\") " pod="openshift-console-operator/console-operator-58897d9998-xjbrz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.625016 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-gd69n" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.625122 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c59c4e2-97ae-494a-b33a-a542266cf233-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.625227 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"a16ed2bfb92b81fce0adc99ac98001bff53b82cfd8c212217dee3d7618503946"} Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.625289 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"44f5ba7b5021740fa36055309a0fed6bca4188708200cd5ea09e952a580b5810"} Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.625331 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/01d95eeb-e421-4fe2-a24a-44c3014f1f6e-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-thcgl\" (UID: \"01d95eeb-e421-4fe2-a24a-44c3014f1f6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.625625 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf612779-d96d-48f1-9958-5f92ac6e3d40-config\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.627347 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c070b67-c2ca-459b-a1b7-813a8833e27e-config\") pod \"route-controller-manager-6576b87f9c-qljmk\" (UID: \"6c070b67-c2ca-459b-a1b7-813a8833e27e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.627946 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72-proxy-tls\") pod \"machine-config-controller-84d6567774-8rrhs\" (UID: \"1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.628218 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/02036666-815d-4282-86f0-1b4163ea7cf9-machine-approver-tls\") pod \"machine-approver-56656f9798-v4kfs\" (UID: \"02036666-815d-4282-86f0-1b4163ea7cf9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.628302 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-oauth-serving-cert\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.628582 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fabbac48-e5db-4057-b9a5-69118c38c667-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-vtkbz\" (UID: \"fabbac48-e5db-4057-b9a5-69118c38c667\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vtkbz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.629102 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b1099221-b48f-4756-b133-0fed16b1e225-proxy-tls\") pod \"machine-config-operator-74547568cd-lrdf2\" (UID: \"b1099221-b48f-4756-b133-0fed16b1e225\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.629650 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5bc3d0cd-eacd-4b25-9acd-853e49db7b47-images\") pod \"machine-api-operator-5694c8668f-fthdq\" (UID: \"5bc3d0cd-eacd-4b25-9acd-853e49db7b47\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.630388 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2c59c4e2-97ae-494a-b33a-a542266cf233-encryption-config\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.630559 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05fe2d3b-0a53-4bbe-b0aa-be07acb159bb-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-qtlw7\" (UID: \"05fe2d3b-0a53-4bbe-b0aa-be07acb159bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.630715 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2c59c4e2-97ae-494a-b33a-a542266cf233-etcd-client\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.630930 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b98dbb3-7986-475d-8028-0879d2fed2af-serving-cert\") pod \"openshift-config-operator-7777fb866f-44l7j\" (UID: \"0b98dbb3-7986-475d-8028-0879d2fed2af\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-44l7j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.630998 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fabbac48-e5db-4057-b9a5-69118c38c667-config\") pod \"kube-apiserver-operator-766d6c64bb-vtkbz\" (UID: \"fabbac48-e5db-4057-b9a5-69118c38c667\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vtkbz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.631030 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2c59c4e2-97ae-494a-b33a-a542266cf233-audit-dir\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.631145 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/0b98dbb3-7986-475d-8028-0879d2fed2af-available-featuregates\") pod \"openshift-config-operator-7777fb866f-44l7j\" (UID: \"0b98dbb3-7986-475d-8028-0879d2fed2af\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-44l7j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.631271 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cf612779-d96d-48f1-9958-5f92ac6e3d40-serving-cert\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.631301 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b81d26a-f92a-4602-89ef-5dd9fb24a32f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-klsxg\" (UID: \"0b81d26a-f92a-4602-89ef-5dd9fb24a32f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-klsxg" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.631498 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05fe2d3b-0a53-4bbe-b0aa-be07acb159bb-service-ca-bundle\") pod \"authentication-operator-69f744f599-qtlw7\" (UID: \"05fe2d3b-0a53-4bbe-b0aa-be07acb159bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.631512 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-trusted-ca-bundle\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.631539 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-9ls2l"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.632485 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/5bc3d0cd-eacd-4b25-9acd-853e49db7b47-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-fthdq\" (UID: \"5bc3d0cd-eacd-4b25-9acd-853e49db7b47\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.632491 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-fhfhr\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.632795 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/cf612779-d96d-48f1-9958-5f92ac6e3d40-etcd-service-ca\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.633038 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/01d95eeb-e421-4fe2-a24a-44c3014f1f6e-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-thcgl\" (UID: \"01d95eeb-e421-4fe2-a24a-44c3014f1f6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.633156 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-client-ca\") pod \"controller-manager-879f6c89f-fhfhr\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.633172 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c38c7edf-ca04-4ea4-b15f-d83dea3f545e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-fl47j\" (UID: \"c38c7edf-ca04-4ea4-b15f-d83dea3f545e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fl47j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.634077 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bbc3437e-8729-46f8-aee2-ad3a4679097f-trusted-ca\") pod \"console-operator-58897d9998-xjbrz\" (UID: \"bbc3437e-8729-46f8-aee2-ad3a4679097f\") " pod="openshift-console-operator/console-operator-58897d9998-xjbrz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.634075 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c59c4e2-97ae-494a-b33a-a542266cf233-serving-cert\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.634268 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c38c7edf-ca04-4ea4-b15f-d83dea3f545e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-fl47j\" (UID: \"c38c7edf-ca04-4ea4-b15f-d83dea3f545e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fl47j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.634312 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-config\") pod \"controller-manager-879f6c89f-fhfhr\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.634678 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05fe2d3b-0a53-4bbe-b0aa-be07acb159bb-serving-cert\") pod \"authentication-operator-69f744f599-qtlw7\" (UID: \"05fe2d3b-0a53-4bbe-b0aa-be07acb159bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.635204 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/de9b95d8-fa65-438c-bef9-2aba46044ec1-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-7x85x\" (UID: \"de9b95d8-fa65-438c-bef9-2aba46044ec1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7x85x" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.635294 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-bp2vx"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.642272 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sjv8t"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.642314 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-cc77j"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.642327 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-46sg8"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.643930 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21db7edd-6db3-4afa-b470-0016a9c5afb9-serving-cert\") pod \"controller-manager-879f6c89f-fhfhr\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.645469 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.646138 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbc3437e-8729-46f8-aee2-ad3a4679097f-config\") pod \"console-operator-58897d9998-xjbrz\" (UID: \"bbc3437e-8729-46f8-aee2-ad3a4679097f\") " pod="openshift-console-operator/console-operator-58897d9998-xjbrz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.648416 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-tdhsc"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.649091 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c070b67-c2ca-459b-a1b7-813a8833e27e-serving-cert\") pod \"route-controller-manager-6576b87f9c-qljmk\" (UID: \"6c070b67-c2ca-459b-a1b7-813a8833e27e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.652700 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-trh7g"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.662277 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.681728 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.687692 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a149fe37-c748-4120-9116-1da4b680d880-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-bgm85\" (UID: \"a149fe37-c748-4120-9116-1da4b680d880\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgm85" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.702935 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.722026 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.722863 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/df73228e-df1f-4e9c-aafe-493bb743f98a-audit-dir\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723007 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723103 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/df73228e-df1f-4e9c-aafe-493bb743f98a-audit-dir\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723107 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6f1e0caa-03ee-4025-9f2a-6941090c178b-metrics-tls\") pod \"dns-operator-744455d44c-x8fgj\" (UID: \"6f1e0caa-03ee-4025-9f2a-6941090c178b\") " pod="openshift-dns-operator/dns-operator-744455d44c-x8fgj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723173 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723265 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c7b67dd6-4025-45ad-98f8-2e7f8ab4b573-profile-collector-cert\") pod \"olm-operator-6b444d44fb-6zdkm\" (UID: \"c7b67dd6-4025-45ad-98f8-2e7f8ab4b573\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723307 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723350 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e9f46a23-4535-4d64-ae76-065435f8f762-bound-sa-token\") pod \"ingress-operator-5b745b69d9-6jlbw\" (UID: \"e9f46a23-4535-4d64-ae76-065435f8f762\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723381 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723403 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e9f46a23-4535-4d64-ae76-065435f8f762-metrics-tls\") pod \"ingress-operator-5b745b69d9-6jlbw\" (UID: \"e9f46a23-4535-4d64-ae76-065435f8f762\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723440 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c7b67dd6-4025-45ad-98f8-2e7f8ab4b573-srv-cert\") pod \"olm-operator-6b444d44fb-6zdkm\" (UID: \"c7b67dd6-4025-45ad-98f8-2e7f8ab4b573\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723466 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2l9z\" (UniqueName: \"kubernetes.io/projected/df73228e-df1f-4e9c-aafe-493bb743f98a-kube-api-access-q2l9z\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723532 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723566 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmpf7\" (UniqueName: \"kubernetes.io/projected/e9f46a23-4535-4d64-ae76-065435f8f762-kube-api-access-lmpf7\") pod \"ingress-operator-5b745b69d9-6jlbw\" (UID: \"e9f46a23-4535-4d64-ae76-065435f8f762\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723589 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrvhw\" (UniqueName: \"kubernetes.io/projected/bc13d9fe-fd58-4d3b-9278-4f3e59da5976-kube-api-access-lrvhw\") pod \"downloads-7954f5f757-grnfc\" (UID: \"bc13d9fe-fd58-4d3b-9278-4f3e59da5976\") " pod="openshift-console/downloads-7954f5f757-grnfc" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723612 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-audit-policies\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723644 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723689 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723746 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hqgh\" (UniqueName: \"kubernetes.io/projected/6f1e0caa-03ee-4025-9f2a-6941090c178b-kube-api-access-6hqgh\") pod \"dns-operator-744455d44c-x8fgj\" (UID: \"6f1e0caa-03ee-4025-9f2a-6941090c178b\") " pod="openshift-dns-operator/dns-operator-744455d44c-x8fgj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723778 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723799 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723822 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723882 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e9f46a23-4535-4d64-ae76-065435f8f762-trusted-ca\") pod \"ingress-operator-5b745b69d9-6jlbw\" (UID: \"e9f46a23-4535-4d64-ae76-065435f8f762\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723911 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723932 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n92xj\" (UniqueName: \"kubernetes.io/projected/c7b67dd6-4025-45ad-98f8-2e7f8ab4b573-kube-api-access-n92xj\") pod \"olm-operator-6b444d44fb-6zdkm\" (UID: \"c7b67dd6-4025-45ad-98f8-2e7f8ab4b573\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.740533 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.761640 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.769069 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6f1e0caa-03ee-4025-9f2a-6941090c178b-metrics-tls\") pod \"dns-operator-744455d44c-x8fgj\" (UID: \"6f1e0caa-03ee-4025-9f2a-6941090c178b\") " pod="openshift-dns-operator/dns-operator-744455d44c-x8fgj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.781384 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.821546 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.841685 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.860930 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.881668 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.902803 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.921535 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.943002 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.961615 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.981572 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.001037 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.022645 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.041412 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.050256 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.062769 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.081950 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.090416 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.100632 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.109139 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.149769 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.155600 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.160713 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.161639 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.166990 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.169275 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.181258 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.187255 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.204899 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.219530 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.241624 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.243266 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.246352 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.261466 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.264994 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-audit-policies\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.282108 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.301527 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.304546 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.321304 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.325359 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.342052 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.361767 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.382600 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.402357 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.421749 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.442133 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.461843 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.481073 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.487703 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c7b67dd6-4025-45ad-98f8-2e7f8ab4b573-profile-collector-cert\") pod \"olm-operator-6b444d44fb-6zdkm\" (UID: \"c7b67dd6-4025-45ad-98f8-2e7f8ab4b573\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.501485 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.520240 4948 request.go:700] Waited for 1.011864135s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/secrets?fieldSelector=metadata.name%3Dpackageserver-service-cert&limit=500&resourceVersion=0 Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.522806 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.541258 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.569195 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.576180 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e9f46a23-4535-4d64-ae76-065435f8f762-trusted-ca\") pod \"ingress-operator-5b745b69d9-6jlbw\" (UID: \"e9f46a23-4535-4d64-ae76-065435f8f762\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.581827 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.601832 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.610083 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e9f46a23-4535-4d64-ae76-065435f8f762-metrics-tls\") pod \"ingress-operator-5b745b69d9-6jlbw\" (UID: \"e9f46a23-4535-4d64-ae76-065435f8f762\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.622708 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.641420 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.662161 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.682058 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.701264 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.708891 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c7b67dd6-4025-45ad-98f8-2e7f8ab4b573-srv-cert\") pod \"olm-operator-6b444d44fb-6zdkm\" (UID: \"c7b67dd6-4025-45ad-98f8-2e7f8ab4b573\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.725130 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.743097 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.761532 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.781351 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.802904 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.821996 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.842073 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.861228 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.881740 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.902601 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.942075 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.952597 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hb9hj\" (UniqueName: \"kubernetes.io/projected/c7c6bf9e-0846-4391-9126-a00d68b2e627-kube-api-access-hb9hj\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.961496 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.982094 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.002050 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.022353 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.041914 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.061591 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.093046 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.102204 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.122838 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.161656 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.180067 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.183586 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.203090 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.223579 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.245624 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.261539 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.281292 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.339059 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fabbac48-e5db-4057-b9a5-69118c38c667-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-vtkbz\" (UID: \"fabbac48-e5db-4057-b9a5-69118c38c667\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vtkbz" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.354105 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkhpt\" (UniqueName: \"kubernetes.io/projected/1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72-kube-api-access-gkhpt\") pod \"machine-config-controller-84d6567774-8rrhs\" (UID: \"1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.375423 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nq9dr\" (UniqueName: \"kubernetes.io/projected/a149fe37-c748-4120-9116-1da4b680d880-kube-api-access-nq9dr\") pod \"control-plane-machine-set-operator-78cbb6b69f-bgm85\" (UID: \"a149fe37-c748-4120-9116-1da4b680d880\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgm85" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.381724 4948 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.391807 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vtkbz" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.394022 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/01d95eeb-e421-4fe2-a24a-44c3014f1f6e-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-thcgl\" (UID: \"01d95eeb-e421-4fe2-a24a-44c3014f1f6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.402645 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.422061 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.463723 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-jfst2"] Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.466369 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mr8r\" (UniqueName: \"kubernetes.io/projected/5bc3d0cd-eacd-4b25-9acd-853e49db7b47-kube-api-access-5mr8r\") pod \"machine-api-operator-5694c8668f-fthdq\" (UID: \"5bc3d0cd-eacd-4b25-9acd-853e49db7b47\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.493057 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94nxl\" (UniqueName: \"kubernetes.io/projected/e029f4ab-6a61-4fcf-9997-b51c1cbe5675-kube-api-access-94nxl\") pod \"kube-storage-version-migrator-operator-b67b599dd-nzj8d\" (UID: \"e029f4ab-6a61-4fcf-9997-b51c1cbe5675\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nzj8d" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.509245 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bt5g\" (UniqueName: \"kubernetes.io/projected/cf612779-d96d-48f1-9958-5f92ac6e3d40-kube-api-access-2bt5g\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.509684 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.514067 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.522782 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mz9sw\" (UniqueName: \"kubernetes.io/projected/de9b95d8-fa65-438c-bef9-2aba46044ec1-kube-api-access-mz9sw\") pod \"cluster-samples-operator-665b6dd947-7x85x\" (UID: \"de9b95d8-fa65-438c-bef9-2aba46044ec1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7x85x" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.529516 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgm85" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.539620 4948 request.go:700] Waited for 1.915562757s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/serviceaccounts/machine-config-operator/token Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.541153 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2v824\" (UniqueName: \"kubernetes.io/projected/01d95eeb-e421-4fe2-a24a-44c3014f1f6e-kube-api-access-2v824\") pod \"cluster-image-registry-operator-dc59b4c8b-thcgl\" (UID: \"01d95eeb-e421-4fe2-a24a-44c3014f1f6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.558237 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgccc\" (UniqueName: \"kubernetes.io/projected/b1099221-b48f-4756-b133-0fed16b1e225-kube-api-access-lgccc\") pod \"machine-config-operator-74547568cd-lrdf2\" (UID: \"b1099221-b48f-4756-b133-0fed16b1e225\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.561513 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.570430 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7x85x" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.578137 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.581487 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.616433 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.648934 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7b5rg\" (UniqueName: \"kubernetes.io/projected/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-kube-api-access-7b5rg\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.655010 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vtkbz"] Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.662369 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.667539 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jfst2" event={"ID":"c7c6bf9e-0846-4391-9126-a00d68b2e627","Type":"ContainerStarted","Data":"810c77e885fc3d0ade548eab4f913ad0847c543d802265ae386fa74cc753a8d0"} Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.668083 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkj4v\" (UniqueName: \"kubernetes.io/projected/05fe2d3b-0a53-4bbe-b0aa-be07acb159bb-kube-api-access-jkj4v\") pod \"authentication-operator-69f744f599-qtlw7\" (UID: \"05fe2d3b-0a53-4bbe-b0aa-be07acb159bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.678911 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.695637 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk7rx\" (UniqueName: \"kubernetes.io/projected/0b81d26a-f92a-4602-89ef-5dd9fb24a32f-kube-api-access-lk7rx\") pod \"openshift-controller-manager-operator-756b6f6bc6-klsxg\" (UID: \"0b81d26a-f92a-4602-89ef-5dd9fb24a32f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-klsxg" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.700429 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9x9v\" (UniqueName: \"kubernetes.io/projected/2c59c4e2-97ae-494a-b33a-a542266cf233-kube-api-access-c9x9v\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:23 crc kubenswrapper[4948]: W0220 08:08:23.714071 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfabbac48_e5db_4057_b9a5_69118c38c667.slice/crio-0078d6f35d4b09153ac5f4519a90739d3b7c1a3332b872561f2b13e7ca2d48f9 WatchSource:0}: Error finding container 0078d6f35d4b09153ac5f4519a90739d3b7c1a3332b872561f2b13e7ca2d48f9: Status 404 returned error can't find the container with id 0078d6f35d4b09153ac5f4519a90739d3b7c1a3332b872561f2b13e7ca2d48f9 Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.716807 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccbnv\" (UniqueName: \"kubernetes.io/projected/0b98dbb3-7986-475d-8028-0879d2fed2af-kube-api-access-ccbnv\") pod \"openshift-config-operator-7777fb866f-44l7j\" (UID: \"0b98dbb3-7986-475d-8028-0879d2fed2af\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-44l7j" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.734956 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xszfj\" (UniqueName: \"kubernetes.io/projected/3f4bc9ec-3eda-4d56-9696-df15c14d2e4a-kube-api-access-xszfj\") pod \"migrator-59844c95c7-cc77j\" (UID: \"3f4bc9ec-3eda-4d56-9696-df15c14d2e4a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-cc77j" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.760552 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhx6j\" (UniqueName: \"kubernetes.io/projected/bbc3437e-8729-46f8-aee2-ad3a4679097f-kube-api-access-lhx6j\") pod \"console-operator-58897d9998-xjbrz\" (UID: \"bbc3437e-8729-46f8-aee2-ad3a4679097f\") " pod="openshift-console-operator/console-operator-58897d9998-xjbrz" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.770093 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nzj8d" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.776837 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.779602 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25nlb\" (UniqueName: \"kubernetes.io/projected/6c070b67-c2ca-459b-a1b7-813a8833e27e-kube-api-access-25nlb\") pod \"route-controller-manager-6576b87f9c-qljmk\" (UID: \"6c070b67-c2ca-459b-a1b7-813a8833e27e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.784926 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-cc77j" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.796599 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.800705 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.802615 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s94wm\" (UniqueName: \"kubernetes.io/projected/02036666-815d-4282-86f0-1b4163ea7cf9-kube-api-access-s94wm\") pod \"machine-approver-56656f9798-v4kfs\" (UID: \"02036666-815d-4282-86f0-1b4163ea7cf9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.815438 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.820472 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-44l7j" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.828505 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9k4m\" (UniqueName: \"kubernetes.io/projected/c38c7edf-ca04-4ea4-b15f-d83dea3f545e-kube-api-access-m9k4m\") pod \"openshift-apiserver-operator-796bbdcf4f-fl47j\" (UID: \"c38c7edf-ca04-4ea4-b15f-d83dea3f545e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fl47j" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.833862 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fl47j" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.836396 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsbfs\" (UniqueName: \"kubernetes.io/projected/21db7edd-6db3-4afa-b470-0016a9c5afb9-kube-api-access-vsbfs\") pod \"controller-manager-879f6c89f-fhfhr\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.874661 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e9f46a23-4535-4d64-ae76-065435f8f762-bound-sa-token\") pod \"ingress-operator-5b745b69d9-6jlbw\" (UID: \"e9f46a23-4535-4d64-ae76-065435f8f762\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.877673 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2l9z\" (UniqueName: \"kubernetes.io/projected/df73228e-df1f-4e9c-aafe-493bb743f98a-kube-api-access-q2l9z\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.892157 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-klsxg" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.906138 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrvhw\" (UniqueName: \"kubernetes.io/projected/bc13d9fe-fd58-4d3b-9278-4f3e59da5976-kube-api-access-lrvhw\") pod \"downloads-7954f5f757-grnfc\" (UID: \"bc13d9fe-fd58-4d3b-9278-4f3e59da5976\") " pod="openshift-console/downloads-7954f5f757-grnfc" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.926788 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmpf7\" (UniqueName: \"kubernetes.io/projected/e9f46a23-4535-4d64-ae76-065435f8f762-kube-api-access-lmpf7\") pod \"ingress-operator-5b745b69d9-6jlbw\" (UID: \"e9f46a23-4535-4d64-ae76-065435f8f762\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.930350 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.937600 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl"] Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.941868 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hqgh\" (UniqueName: \"kubernetes.io/projected/6f1e0caa-03ee-4025-9f2a-6941090c178b-kube-api-access-6hqgh\") pod \"dns-operator-744455d44c-x8fgj\" (UID: \"6f1e0caa-03ee-4025-9f2a-6941090c178b\") " pod="openshift-dns-operator/dns-operator-744455d44c-x8fgj" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.964851 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n92xj\" (UniqueName: \"kubernetes.io/projected/c7b67dd6-4025-45ad-98f8-2e7f8ab4b573-kube-api-access-n92xj\") pod \"olm-operator-6b444d44fb-6zdkm\" (UID: \"c7b67dd6-4025-45ad-98f8-2e7f8ab4b573\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.975878 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.992675 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-fthdq"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.008504 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-xjbrz" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.010688 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.011195 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qfgtj"] Feb 20 08:08:24 crc kubenswrapper[4948]: W0220 08:08:24.023658 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod01d95eeb_e421_4fe2_a24a_44c3014f1f6e.slice/crio-fd9df30c5d63b531a8098b8313e0aead45e05e65f20b2d3a9242e2f927d8db1f WatchSource:0}: Error finding container fd9df30c5d63b531a8098b8313e0aead45e05e65f20b2d3a9242e2f927d8db1f: Status 404 returned error can't find the container with id fd9df30c5d63b531a8098b8313e0aead45e05e65f20b2d3a9242e2f927d8db1f Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.055623 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7x85x"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.071774 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgm85"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088111 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/308389f1-6962-45be-87b0-1b061caa2c50-signing-key\") pod \"service-ca-9c57cc56f-rtnmx\" (UID: \"308389f1-6962-45be-87b0-1b061caa2c50\") " pod="openshift-service-ca/service-ca-9c57cc56f-rtnmx" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088450 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmt4n\" (UniqueName: \"kubernetes.io/projected/912a3267-581f-47b1-873c-b0e4d8dd3768-kube-api-access-zmt4n\") pod \"catalog-operator-68c6474976-dgw2h\" (UID: \"912a3267-581f-47b1-873c-b0e4d8dd3768\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088494 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/fb7c802e-5175-4b0b-8b70-91efb1c83fa1-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-46sg8\" (UID: \"fb7c802e-5175-4b0b-8b70-91efb1c83fa1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-46sg8" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088536 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/99fafd36-91d7-45b8-a321-59307d79e5ce-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-5kfk4\" (UID: \"99fafd36-91d7-45b8-a321-59307d79e5ce\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5kfk4" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088598 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55q6w\" (UniqueName: \"kubernetes.io/projected/95a270e5-7a08-4238-a65d-96986f2d2c39-kube-api-access-55q6w\") pod \"marketplace-operator-79b997595-sjv8t\" (UID: \"95a270e5-7a08-4238-a65d-96986f2d2c39\") " pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088617 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npgsj\" (UniqueName: \"kubernetes.io/projected/000483d2-8b2d-4403-9652-d4d86d55f7b5-kube-api-access-npgsj\") pod \"multus-admission-controller-857f4d67dd-bcsmj\" (UID: \"000483d2-8b2d-4403-9652-d4d86d55f7b5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-bcsmj" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088672 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4b3348f9-d9b9-4efb-9eb8-877f58b0d59c-metrics-certs\") pod \"router-default-5444994796-sc5pm\" (UID: \"4b3348f9-d9b9-4efb-9eb8-877f58b0d59c\") " pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088689 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fj9kq\" (UniqueName: \"kubernetes.io/projected/4b3348f9-d9b9-4efb-9eb8-877f58b0d59c-kube-api-access-fj9kq\") pod \"router-default-5444994796-sc5pm\" (UID: \"4b3348f9-d9b9-4efb-9eb8-877f58b0d59c\") " pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088705 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99fafd36-91d7-45b8-a321-59307d79e5ce-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-5kfk4\" (UID: \"99fafd36-91d7-45b8-a321-59307d79e5ce\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5kfk4" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088724 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40656b16-31ff-4065-b87c-5678308e9fb4-config\") pod \"service-ca-operator-777779d784-tdhsc\" (UID: \"40656b16-31ff-4065-b87c-5678308e9fb4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tdhsc" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088744 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/57c455f7-aa9c-405c-bee3-89726b84f5db-installation-pull-secrets\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088778 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99fafd36-91d7-45b8-a321-59307d79e5ce-config\") pod \"kube-controller-manager-operator-78b949d7b-5kfk4\" (UID: \"99fafd36-91d7-45b8-a321-59307d79e5ce\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5kfk4" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088806 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a102d1fa-69a5-4826-ac40-e07ab4087558-apiservice-cert\") pod \"packageserver-d55dfcdfc-dkdgm\" (UID: \"a102d1fa-69a5-4826-ac40-e07ab4087558\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088825 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/57c455f7-aa9c-405c-bee3-89726b84f5db-registry-certificates\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088860 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qgskm\" (UID: \"e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qgskm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088874 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/912a3267-581f-47b1-873c-b0e4d8dd3768-srv-cert\") pod \"catalog-operator-68c6474976-dgw2h\" (UID: \"912a3267-581f-47b1-873c-b0e4d8dd3768\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088915 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsw92\" (UniqueName: \"kubernetes.io/projected/fb7c802e-5175-4b0b-8b70-91efb1c83fa1-kube-api-access-nsw92\") pod \"package-server-manager-789f6589d5-46sg8\" (UID: \"fb7c802e-5175-4b0b-8b70-91efb1c83fa1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-46sg8" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088935 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vg4v\" (UniqueName: \"kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-kube-api-access-2vg4v\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089009 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/4b3348f9-d9b9-4efb-9eb8-877f58b0d59c-stats-auth\") pod \"router-default-5444994796-sc5pm\" (UID: \"4b3348f9-d9b9-4efb-9eb8-877f58b0d59c\") " pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089044 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/95a270e5-7a08-4238-a65d-96986f2d2c39-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sjv8t\" (UID: \"95a270e5-7a08-4238-a65d-96986f2d2c39\") " pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089095 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089114 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qgskm\" (UID: \"e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qgskm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089131 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/4b3348f9-d9b9-4efb-9eb8-877f58b0d59c-default-certificate\") pod \"router-default-5444994796-sc5pm\" (UID: \"4b3348f9-d9b9-4efb-9eb8-877f58b0d59c\") " pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089164 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjwcj\" (UniqueName: \"kubernetes.io/projected/308389f1-6962-45be-87b0-1b061caa2c50-kube-api-access-vjwcj\") pod \"service-ca-9c57cc56f-rtnmx\" (UID: \"308389f1-6962-45be-87b0-1b061caa2c50\") " pod="openshift-service-ca/service-ca-9c57cc56f-rtnmx" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089179 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qgskm\" (UID: \"e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qgskm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089195 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/95a270e5-7a08-4238-a65d-96986f2d2c39-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sjv8t\" (UID: \"95a270e5-7a08-4238-a65d-96986f2d2c39\") " pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089211 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f77016a6-cd35-49df-b6b5-65b4858b41c9-config-volume\") pod \"collect-profiles-29526240-lwsqk\" (UID: \"f77016a6-cd35-49df-b6b5-65b4858b41c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089229 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/000483d2-8b2d-4403-9652-d4d86d55f7b5-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-bcsmj\" (UID: \"000483d2-8b2d-4403-9652-d4d86d55f7b5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-bcsmj" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089253 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-registry-tls\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089287 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svs7x\" (UniqueName: \"kubernetes.io/projected/a102d1fa-69a5-4826-ac40-e07ab4087558-kube-api-access-svs7x\") pod \"packageserver-d55dfcdfc-dkdgm\" (UID: \"a102d1fa-69a5-4826-ac40-e07ab4087558\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089303 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/912a3267-581f-47b1-873c-b0e4d8dd3768-profile-collector-cert\") pod \"catalog-operator-68c6474976-dgw2h\" (UID: \"912a3267-581f-47b1-873c-b0e4d8dd3768\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089321 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ql4r\" (UniqueName: \"kubernetes.io/projected/f77016a6-cd35-49df-b6b5-65b4858b41c9-kube-api-access-2ql4r\") pod \"collect-profiles-29526240-lwsqk\" (UID: \"f77016a6-cd35-49df-b6b5-65b4858b41c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089358 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/57c455f7-aa9c-405c-bee3-89726b84f5db-trusted-ca\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089406 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-bound-sa-token\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089423 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a102d1fa-69a5-4826-ac40-e07ab4087558-tmpfs\") pod \"packageserver-d55dfcdfc-dkdgm\" (UID: \"a102d1fa-69a5-4826-ac40-e07ab4087558\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089457 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/57c455f7-aa9c-405c-bee3-89726b84f5db-ca-trust-extracted\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089473 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/308389f1-6962-45be-87b0-1b061caa2c50-signing-cabundle\") pod \"service-ca-9c57cc56f-rtnmx\" (UID: \"308389f1-6962-45be-87b0-1b061caa2c50\") " pod="openshift-service-ca/service-ca-9c57cc56f-rtnmx" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089500 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a102d1fa-69a5-4826-ac40-e07ab4087558-webhook-cert\") pod \"packageserver-d55dfcdfc-dkdgm\" (UID: \"a102d1fa-69a5-4826-ac40-e07ab4087558\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089524 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b3348f9-d9b9-4efb-9eb8-877f58b0d59c-service-ca-bundle\") pod \"router-default-5444994796-sc5pm\" (UID: \"4b3348f9-d9b9-4efb-9eb8-877f58b0d59c\") " pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089547 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ng7x4\" (UniqueName: \"kubernetes.io/projected/40656b16-31ff-4065-b87c-5678308e9fb4-kube-api-access-ng7x4\") pod \"service-ca-operator-777779d784-tdhsc\" (UID: \"40656b16-31ff-4065-b87c-5678308e9fb4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tdhsc" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089666 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40656b16-31ff-4065-b87c-5678308e9fb4-serving-cert\") pod \"service-ca-operator-777779d784-tdhsc\" (UID: \"40656b16-31ff-4065-b87c-5678308e9fb4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tdhsc" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089719 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f77016a6-cd35-49df-b6b5-65b4858b41c9-secret-volume\") pod \"collect-profiles-29526240-lwsqk\" (UID: \"f77016a6-cd35-49df-b6b5-65b4858b41c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" Feb 20 08:08:24 crc kubenswrapper[4948]: E0220 08:08:24.091957 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:24.591937614 +0000 UTC m=+153.566432434 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.136920 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-x8fgj" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.168285 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-grnfc" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.169477 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.190412 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:24 crc kubenswrapper[4948]: E0220 08:08:24.190659 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:24.690627935 +0000 UTC m=+153.665122755 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.190992 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a102d1fa-69a5-4826-ac40-e07ab4087558-webhook-cert\") pod \"packageserver-d55dfcdfc-dkdgm\" (UID: \"a102d1fa-69a5-4826-ac40-e07ab4087558\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191021 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b3348f9-d9b9-4efb-9eb8-877f58b0d59c-service-ca-bundle\") pod \"router-default-5444994796-sc5pm\" (UID: \"4b3348f9-d9b9-4efb-9eb8-877f58b0d59c\") " pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191043 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ng7x4\" (UniqueName: \"kubernetes.io/projected/40656b16-31ff-4065-b87c-5678308e9fb4-kube-api-access-ng7x4\") pod \"service-ca-operator-777779d784-tdhsc\" (UID: \"40656b16-31ff-4065-b87c-5678308e9fb4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tdhsc" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191067 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40656b16-31ff-4065-b87c-5678308e9fb4-serving-cert\") pod \"service-ca-operator-777779d784-tdhsc\" (UID: \"40656b16-31ff-4065-b87c-5678308e9fb4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tdhsc" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191111 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f77016a6-cd35-49df-b6b5-65b4858b41c9-secret-volume\") pod \"collect-profiles-29526240-lwsqk\" (UID: \"f77016a6-cd35-49df-b6b5-65b4858b41c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191133 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/308389f1-6962-45be-87b0-1b061caa2c50-signing-key\") pod \"service-ca-9c57cc56f-rtnmx\" (UID: \"308389f1-6962-45be-87b0-1b061caa2c50\") " pod="openshift-service-ca/service-ca-9c57cc56f-rtnmx" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191169 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pd7s\" (UniqueName: \"kubernetes.io/projected/9dbe3199-893b-4d67-a447-fa3223c321e3-kube-api-access-9pd7s\") pod \"ingress-canary-bp2vx\" (UID: \"9dbe3199-893b-4d67-a447-fa3223c321e3\") " pod="openshift-ingress-canary/ingress-canary-bp2vx" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191202 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmt4n\" (UniqueName: \"kubernetes.io/projected/912a3267-581f-47b1-873c-b0e4d8dd3768-kube-api-access-zmt4n\") pod \"catalog-operator-68c6474976-dgw2h\" (UID: \"912a3267-581f-47b1-873c-b0e4d8dd3768\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191243 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/fb7c802e-5175-4b0b-8b70-91efb1c83fa1-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-46sg8\" (UID: \"fb7c802e-5175-4b0b-8b70-91efb1c83fa1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-46sg8" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191267 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156-node-bootstrap-token\") pod \"machine-config-server-gd69n\" (UID: \"bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156\") " pod="openshift-machine-config-operator/machine-config-server-gd69n" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191297 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/99fafd36-91d7-45b8-a321-59307d79e5ce-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-5kfk4\" (UID: \"99fafd36-91d7-45b8-a321-59307d79e5ce\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5kfk4" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191340 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55q6w\" (UniqueName: \"kubernetes.io/projected/95a270e5-7a08-4238-a65d-96986f2d2c39-kube-api-access-55q6w\") pod \"marketplace-operator-79b997595-sjv8t\" (UID: \"95a270e5-7a08-4238-a65d-96986f2d2c39\") " pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191371 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npgsj\" (UniqueName: \"kubernetes.io/projected/000483d2-8b2d-4403-9652-d4d86d55f7b5-kube-api-access-npgsj\") pod \"multus-admission-controller-857f4d67dd-bcsmj\" (UID: \"000483d2-8b2d-4403-9652-d4d86d55f7b5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-bcsmj" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191395 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qv78n\" (UniqueName: \"kubernetes.io/projected/06e36dcf-9973-4369-8dc0-735cc42a68aa-kube-api-access-qv78n\") pod \"dns-default-9ls2l\" (UID: \"06e36dcf-9973-4369-8dc0-735cc42a68aa\") " pod="openshift-dns/dns-default-9ls2l" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191428 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/13b4eba0-42e7-448a-998c-d7aaeffc9218-csi-data-dir\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191481 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4b3348f9-d9b9-4efb-9eb8-877f58b0d59c-metrics-certs\") pod \"router-default-5444994796-sc5pm\" (UID: \"4b3348f9-d9b9-4efb-9eb8-877f58b0d59c\") " pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191501 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/06e36dcf-9973-4369-8dc0-735cc42a68aa-config-volume\") pod \"dns-default-9ls2l\" (UID: \"06e36dcf-9973-4369-8dc0-735cc42a68aa\") " pod="openshift-dns/dns-default-9ls2l" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191522 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fj9kq\" (UniqueName: \"kubernetes.io/projected/4b3348f9-d9b9-4efb-9eb8-877f58b0d59c-kube-api-access-fj9kq\") pod \"router-default-5444994796-sc5pm\" (UID: \"4b3348f9-d9b9-4efb-9eb8-877f58b0d59c\") " pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191545 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99fafd36-91d7-45b8-a321-59307d79e5ce-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-5kfk4\" (UID: \"99fafd36-91d7-45b8-a321-59307d79e5ce\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5kfk4" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191563 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40656b16-31ff-4065-b87c-5678308e9fb4-config\") pod \"service-ca-operator-777779d784-tdhsc\" (UID: \"40656b16-31ff-4065-b87c-5678308e9fb4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tdhsc" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191595 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/57c455f7-aa9c-405c-bee3-89726b84f5db-installation-pull-secrets\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191615 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99fafd36-91d7-45b8-a321-59307d79e5ce-config\") pod \"kube-controller-manager-operator-78b949d7b-5kfk4\" (UID: \"99fafd36-91d7-45b8-a321-59307d79e5ce\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5kfk4" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191633 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/06e36dcf-9973-4369-8dc0-735cc42a68aa-metrics-tls\") pod \"dns-default-9ls2l\" (UID: \"06e36dcf-9973-4369-8dc0-735cc42a68aa\") " pod="openshift-dns/dns-default-9ls2l" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191670 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a102d1fa-69a5-4826-ac40-e07ab4087558-apiservice-cert\") pod \"packageserver-d55dfcdfc-dkdgm\" (UID: \"a102d1fa-69a5-4826-ac40-e07ab4087558\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191711 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/57c455f7-aa9c-405c-bee3-89726b84f5db-registry-certificates\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191742 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qgskm\" (UID: \"e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qgskm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191761 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/912a3267-581f-47b1-873c-b0e4d8dd3768-srv-cert\") pod \"catalog-operator-68c6474976-dgw2h\" (UID: \"912a3267-581f-47b1-873c-b0e4d8dd3768\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191779 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsw92\" (UniqueName: \"kubernetes.io/projected/fb7c802e-5175-4b0b-8b70-91efb1c83fa1-kube-api-access-nsw92\") pod \"package-server-manager-789f6589d5-46sg8\" (UID: \"fb7c802e-5175-4b0b-8b70-91efb1c83fa1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-46sg8" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191823 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vg4v\" (UniqueName: \"kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-kube-api-access-2vg4v\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191905 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/4b3348f9-d9b9-4efb-9eb8-877f58b0d59c-stats-auth\") pod \"router-default-5444994796-sc5pm\" (UID: \"4b3348f9-d9b9-4efb-9eb8-877f58b0d59c\") " pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191934 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/95a270e5-7a08-4238-a65d-96986f2d2c39-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sjv8t\" (UID: \"95a270e5-7a08-4238-a65d-96986f2d2c39\") " pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191953 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156-certs\") pod \"machine-config-server-gd69n\" (UID: \"bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156\") " pod="openshift-machine-config-operator/machine-config-server-gd69n" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.192027 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.192047 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qgskm\" (UID: \"e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qgskm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.192066 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/4b3348f9-d9b9-4efb-9eb8-877f58b0d59c-default-certificate\") pod \"router-default-5444994796-sc5pm\" (UID: \"4b3348f9-d9b9-4efb-9eb8-877f58b0d59c\") " pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.192100 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjwcj\" (UniqueName: \"kubernetes.io/projected/308389f1-6962-45be-87b0-1b061caa2c50-kube-api-access-vjwcj\") pod \"service-ca-9c57cc56f-rtnmx\" (UID: \"308389f1-6962-45be-87b0-1b061caa2c50\") " pod="openshift-service-ca/service-ca-9c57cc56f-rtnmx" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.192118 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/13b4eba0-42e7-448a-998c-d7aaeffc9218-socket-dir\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.192137 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qgskm\" (UID: \"e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qgskm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.192155 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtrdt\" (UniqueName: \"kubernetes.io/projected/bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156-kube-api-access-qtrdt\") pod \"machine-config-server-gd69n\" (UID: \"bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156\") " pod="openshift-machine-config-operator/machine-config-server-gd69n" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.192210 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/95a270e5-7a08-4238-a65d-96986f2d2c39-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sjv8t\" (UID: \"95a270e5-7a08-4238-a65d-96986f2d2c39\") " pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.192841 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f77016a6-cd35-49df-b6b5-65b4858b41c9-config-volume\") pod \"collect-profiles-29526240-lwsqk\" (UID: \"f77016a6-cd35-49df-b6b5-65b4858b41c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.192872 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/000483d2-8b2d-4403-9652-d4d86d55f7b5-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-bcsmj\" (UID: \"000483d2-8b2d-4403-9652-d4d86d55f7b5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-bcsmj" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.192914 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-registry-tls\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.192962 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svs7x\" (UniqueName: \"kubernetes.io/projected/a102d1fa-69a5-4826-ac40-e07ab4087558-kube-api-access-svs7x\") pod \"packageserver-d55dfcdfc-dkdgm\" (UID: \"a102d1fa-69a5-4826-ac40-e07ab4087558\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.192992 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/912a3267-581f-47b1-873c-b0e4d8dd3768-profile-collector-cert\") pod \"catalog-operator-68c6474976-dgw2h\" (UID: \"912a3267-581f-47b1-873c-b0e4d8dd3768\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.193021 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/13b4eba0-42e7-448a-998c-d7aaeffc9218-plugins-dir\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.193049 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/13b4eba0-42e7-448a-998c-d7aaeffc9218-mountpoint-dir\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.193079 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ql4r\" (UniqueName: \"kubernetes.io/projected/f77016a6-cd35-49df-b6b5-65b4858b41c9-kube-api-access-2ql4r\") pod \"collect-profiles-29526240-lwsqk\" (UID: \"f77016a6-cd35-49df-b6b5-65b4858b41c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.193097 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9dbe3199-893b-4d67-a447-fa3223c321e3-cert\") pod \"ingress-canary-bp2vx\" (UID: \"9dbe3199-893b-4d67-a447-fa3223c321e3\") " pod="openshift-ingress-canary/ingress-canary-bp2vx" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.193115 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/13b4eba0-42e7-448a-998c-d7aaeffc9218-registration-dir\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.193132 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5jbw\" (UniqueName: \"kubernetes.io/projected/13b4eba0-42e7-448a-998c-d7aaeffc9218-kube-api-access-v5jbw\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.193154 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/57c455f7-aa9c-405c-bee3-89726b84f5db-trusted-ca\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.193212 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-bound-sa-token\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.193228 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a102d1fa-69a5-4826-ac40-e07ab4087558-tmpfs\") pod \"packageserver-d55dfcdfc-dkdgm\" (UID: \"a102d1fa-69a5-4826-ac40-e07ab4087558\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.193256 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/57c455f7-aa9c-405c-bee3-89726b84f5db-ca-trust-extracted\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.193276 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/308389f1-6962-45be-87b0-1b061caa2c50-signing-cabundle\") pod \"service-ca-9c57cc56f-rtnmx\" (UID: \"308389f1-6962-45be-87b0-1b061caa2c50\") " pod="openshift-service-ca/service-ca-9c57cc56f-rtnmx" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.194544 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b3348f9-d9b9-4efb-9eb8-877f58b0d59c-service-ca-bundle\") pod \"router-default-5444994796-sc5pm\" (UID: \"4b3348f9-d9b9-4efb-9eb8-877f58b0d59c\") " pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.195638 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/57c455f7-aa9c-405c-bee3-89726b84f5db-registry-certificates\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: E0220 08:08:24.196146 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:24.69613101 +0000 UTC m=+153.670625830 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.197790 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/57c455f7-aa9c-405c-bee3-89726b84f5db-trusted-ca\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.198485 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a102d1fa-69a5-4826-ac40-e07ab4087558-tmpfs\") pod \"packageserver-d55dfcdfc-dkdgm\" (UID: \"a102d1fa-69a5-4826-ac40-e07ab4087558\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.198723 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/57c455f7-aa9c-405c-bee3-89726b84f5db-ca-trust-extracted\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.199284 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/308389f1-6962-45be-87b0-1b061caa2c50-signing-cabundle\") pod \"service-ca-9c57cc56f-rtnmx\" (UID: \"308389f1-6962-45be-87b0-1b061caa2c50\") " pod="openshift-service-ca/service-ca-9c57cc56f-rtnmx" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.201369 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40656b16-31ff-4065-b87c-5678308e9fb4-config\") pod \"service-ca-operator-777779d784-tdhsc\" (UID: \"40656b16-31ff-4065-b87c-5678308e9fb4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tdhsc" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.201818 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40656b16-31ff-4065-b87c-5678308e9fb4-serving-cert\") pod \"service-ca-operator-777779d784-tdhsc\" (UID: \"40656b16-31ff-4065-b87c-5678308e9fb4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tdhsc" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.202171 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/95a270e5-7a08-4238-a65d-96986f2d2c39-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sjv8t\" (UID: \"95a270e5-7a08-4238-a65d-96986f2d2c39\") " pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.203825 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/95a270e5-7a08-4238-a65d-96986f2d2c39-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sjv8t\" (UID: \"95a270e5-7a08-4238-a65d-96986f2d2c39\") " pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.203941 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qgskm\" (UID: \"e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qgskm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.204248 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a102d1fa-69a5-4826-ac40-e07ab4087558-apiservice-cert\") pod \"packageserver-d55dfcdfc-dkdgm\" (UID: \"a102d1fa-69a5-4826-ac40-e07ab4087558\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.205388 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/4b3348f9-d9b9-4efb-9eb8-877f58b0d59c-stats-auth\") pod \"router-default-5444994796-sc5pm\" (UID: \"4b3348f9-d9b9-4efb-9eb8-877f58b0d59c\") " pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.206505 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f77016a6-cd35-49df-b6b5-65b4858b41c9-config-volume\") pod \"collect-profiles-29526240-lwsqk\" (UID: \"f77016a6-cd35-49df-b6b5-65b4858b41c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.208446 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/57c455f7-aa9c-405c-bee3-89726b84f5db-installation-pull-secrets\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.210617 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/308389f1-6962-45be-87b0-1b061caa2c50-signing-key\") pod \"service-ca-9c57cc56f-rtnmx\" (UID: \"308389f1-6962-45be-87b0-1b061caa2c50\") " pod="openshift-service-ca/service-ca-9c57cc56f-rtnmx" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.211635 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/912a3267-581f-47b1-873c-b0e4d8dd3768-srv-cert\") pod \"catalog-operator-68c6474976-dgw2h\" (UID: \"912a3267-581f-47b1-873c-b0e4d8dd3768\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.212007 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/fb7c802e-5175-4b0b-8b70-91efb1c83fa1-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-46sg8\" (UID: \"fb7c802e-5175-4b0b-8b70-91efb1c83fa1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-46sg8" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.212489 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a102d1fa-69a5-4826-ac40-e07ab4087558-webhook-cert\") pod \"packageserver-d55dfcdfc-dkdgm\" (UID: \"a102d1fa-69a5-4826-ac40-e07ab4087558\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.212597 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/4b3348f9-d9b9-4efb-9eb8-877f58b0d59c-default-certificate\") pod \"router-default-5444994796-sc5pm\" (UID: \"4b3348f9-d9b9-4efb-9eb8-877f58b0d59c\") " pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.213186 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/912a3267-581f-47b1-873c-b0e4d8dd3768-profile-collector-cert\") pod \"catalog-operator-68c6474976-dgw2h\" (UID: \"912a3267-581f-47b1-873c-b0e4d8dd3768\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.213480 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99fafd36-91d7-45b8-a321-59307d79e5ce-config\") pod \"kube-controller-manager-operator-78b949d7b-5kfk4\" (UID: \"99fafd36-91d7-45b8-a321-59307d79e5ce\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5kfk4" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.214258 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4b3348f9-d9b9-4efb-9eb8-877f58b0d59c-metrics-certs\") pod \"router-default-5444994796-sc5pm\" (UID: \"4b3348f9-d9b9-4efb-9eb8-877f58b0d59c\") " pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.214346 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99fafd36-91d7-45b8-a321-59307d79e5ce-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-5kfk4\" (UID: \"99fafd36-91d7-45b8-a321-59307d79e5ce\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5kfk4" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.219221 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.227768 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-registry-tls\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.227990 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/000483d2-8b2d-4403-9652-d4d86d55f7b5-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-bcsmj\" (UID: \"000483d2-8b2d-4403-9652-d4d86d55f7b5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-bcsmj" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.242716 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsw92\" (UniqueName: \"kubernetes.io/projected/fb7c802e-5175-4b0b-8b70-91efb1c83fa1-kube-api-access-nsw92\") pod \"package-server-manager-789f6589d5-46sg8\" (UID: \"fb7c802e-5175-4b0b-8b70-91efb1c83fa1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-46sg8" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.247229 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qgskm\" (UID: \"e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qgskm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.247389 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f77016a6-cd35-49df-b6b5-65b4858b41c9-secret-volume\") pod \"collect-profiles-29526240-lwsqk\" (UID: \"f77016a6-cd35-49df-b6b5-65b4858b41c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.269206 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-46sg8" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.273936 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.277426 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vg4v\" (UniqueName: \"kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-kube-api-access-2vg4v\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.289987 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nzj8d"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.291140 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ng7x4\" (UniqueName: \"kubernetes.io/projected/40656b16-31ff-4065-b87c-5678308e9fb4-kube-api-access-ng7x4\") pod \"service-ca-operator-777779d784-tdhsc\" (UID: \"40656b16-31ff-4065-b87c-5678308e9fb4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tdhsc" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.292132 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-tdhsc" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.293586 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:24 crc kubenswrapper[4948]: E0220 08:08:24.293773 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:24.793756464 +0000 UTC m=+153.768251284 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.293799 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pd7s\" (UniqueName: \"kubernetes.io/projected/9dbe3199-893b-4d67-a447-fa3223c321e3-kube-api-access-9pd7s\") pod \"ingress-canary-bp2vx\" (UID: \"9dbe3199-893b-4d67-a447-fa3223c321e3\") " pod="openshift-ingress-canary/ingress-canary-bp2vx" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.293831 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156-node-bootstrap-token\") pod \"machine-config-server-gd69n\" (UID: \"bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156\") " pod="openshift-machine-config-operator/machine-config-server-gd69n" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.293867 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qv78n\" (UniqueName: \"kubernetes.io/projected/06e36dcf-9973-4369-8dc0-735cc42a68aa-kube-api-access-qv78n\") pod \"dns-default-9ls2l\" (UID: \"06e36dcf-9973-4369-8dc0-735cc42a68aa\") " pod="openshift-dns/dns-default-9ls2l" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.293886 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/13b4eba0-42e7-448a-998c-d7aaeffc9218-csi-data-dir\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.294643 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/06e36dcf-9973-4369-8dc0-735cc42a68aa-config-volume\") pod \"dns-default-9ls2l\" (UID: \"06e36dcf-9973-4369-8dc0-735cc42a68aa\") " pod="openshift-dns/dns-default-9ls2l" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.294678 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/06e36dcf-9973-4369-8dc0-735cc42a68aa-metrics-tls\") pod \"dns-default-9ls2l\" (UID: \"06e36dcf-9973-4369-8dc0-735cc42a68aa\") " pod="openshift-dns/dns-default-9ls2l" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.294715 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156-certs\") pod \"machine-config-server-gd69n\" (UID: \"bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156\") " pod="openshift-machine-config-operator/machine-config-server-gd69n" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.294739 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.294767 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/13b4eba0-42e7-448a-998c-d7aaeffc9218-socket-dir\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.294786 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtrdt\" (UniqueName: \"kubernetes.io/projected/bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156-kube-api-access-qtrdt\") pod \"machine-config-server-gd69n\" (UID: \"bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156\") " pod="openshift-machine-config-operator/machine-config-server-gd69n" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.294819 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/13b4eba0-42e7-448a-998c-d7aaeffc9218-plugins-dir\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.294845 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/13b4eba0-42e7-448a-998c-d7aaeffc9218-mountpoint-dir\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.294878 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9dbe3199-893b-4d67-a447-fa3223c321e3-cert\") pod \"ingress-canary-bp2vx\" (UID: \"9dbe3199-893b-4d67-a447-fa3223c321e3\") " pod="openshift-ingress-canary/ingress-canary-bp2vx" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.294900 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/13b4eba0-42e7-448a-998c-d7aaeffc9218-registration-dir\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.294920 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5jbw\" (UniqueName: \"kubernetes.io/projected/13b4eba0-42e7-448a-998c-d7aaeffc9218-kube-api-access-v5jbw\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.294679 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-qtlw7"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.296849 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/13b4eba0-42e7-448a-998c-d7aaeffc9218-plugins-dir\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: E0220 08:08:24.297236 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:24.797217199 +0000 UTC m=+153.771712019 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.297312 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/13b4eba0-42e7-448a-998c-d7aaeffc9218-mountpoint-dir\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.297448 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/13b4eba0-42e7-448a-998c-d7aaeffc9218-socket-dir\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.297455 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/13b4eba0-42e7-448a-998c-d7aaeffc9218-csi-data-dir\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.297496 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/13b4eba0-42e7-448a-998c-d7aaeffc9218-registration-dir\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.298052 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/06e36dcf-9973-4369-8dc0-735cc42a68aa-config-volume\") pod \"dns-default-9ls2l\" (UID: \"06e36dcf-9973-4369-8dc0-735cc42a68aa\") " pod="openshift-dns/dns-default-9ls2l" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.298228 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156-certs\") pod \"machine-config-server-gd69n\" (UID: \"bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156\") " pod="openshift-machine-config-operator/machine-config-server-gd69n" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.299653 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qgskm\" (UID: \"e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qgskm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.299788 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/06e36dcf-9973-4369-8dc0-735cc42a68aa-metrics-tls\") pod \"dns-default-9ls2l\" (UID: \"06e36dcf-9973-4369-8dc0-735cc42a68aa\") " pod="openshift-dns/dns-default-9ls2l" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.300718 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156-node-bootstrap-token\") pod \"machine-config-server-gd69n\" (UID: \"bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156\") " pod="openshift-machine-config-operator/machine-config-server-gd69n" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.311519 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9dbe3199-893b-4d67-a447-fa3223c321e3-cert\") pod \"ingress-canary-bp2vx\" (UID: \"9dbe3199-893b-4d67-a447-fa3223c321e3\") " pod="openshift-ingress-canary/ingress-canary-bp2vx" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.315307 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fj9kq\" (UniqueName: \"kubernetes.io/projected/4b3348f9-d9b9-4efb-9eb8-877f58b0d59c-kube-api-access-fj9kq\") pod \"router-default-5444994796-sc5pm\" (UID: \"4b3348f9-d9b9-4efb-9eb8-877f58b0d59c\") " pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.352318 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svs7x\" (UniqueName: \"kubernetes.io/projected/a102d1fa-69a5-4826-ac40-e07ab4087558-kube-api-access-svs7x\") pod \"packageserver-d55dfcdfc-dkdgm\" (UID: \"a102d1fa-69a5-4826-ac40-e07ab4087558\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.371095 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-bound-sa-token\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.390495 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/99fafd36-91d7-45b8-a321-59307d79e5ce-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-5kfk4\" (UID: \"99fafd36-91d7-45b8-a321-59307d79e5ce\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5kfk4" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.396066 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:24 crc kubenswrapper[4948]: E0220 08:08:24.396446 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:24.896430981 +0000 UTC m=+153.870925801 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.401702 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjwcj\" (UniqueName: \"kubernetes.io/projected/308389f1-6962-45be-87b0-1b061caa2c50-kube-api-access-vjwcj\") pod \"service-ca-9c57cc56f-rtnmx\" (UID: \"308389f1-6962-45be-87b0-1b061caa2c50\") " pod="openshift-service-ca/service-ca-9c57cc56f-rtnmx" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.432044 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ql4r\" (UniqueName: \"kubernetes.io/projected/f77016a6-cd35-49df-b6b5-65b4858b41c9-kube-api-access-2ql4r\") pod \"collect-profiles-29526240-lwsqk\" (UID: \"f77016a6-cd35-49df-b6b5-65b4858b41c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.444571 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55q6w\" (UniqueName: \"kubernetes.io/projected/95a270e5-7a08-4238-a65d-96986f2d2c39-kube-api-access-55q6w\") pod \"marketplace-operator-79b997595-sjv8t\" (UID: \"95a270e5-7a08-4238-a65d-96986f2d2c39\") " pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.445382 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qgskm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.458893 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.472241 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npgsj\" (UniqueName: \"kubernetes.io/projected/000483d2-8b2d-4403-9652-d4d86d55f7b5-kube-api-access-npgsj\") pod \"multus-admission-controller-857f4d67dd-bcsmj\" (UID: \"000483d2-8b2d-4403-9652-d4d86d55f7b5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-bcsmj" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.495882 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmt4n\" (UniqueName: \"kubernetes.io/projected/912a3267-581f-47b1-873c-b0e4d8dd3768-kube-api-access-zmt4n\") pod \"catalog-operator-68c6474976-dgw2h\" (UID: \"912a3267-581f-47b1-873c-b0e4d8dd3768\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.497248 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: E0220 08:08:24.497720 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:24.997704094 +0000 UTC m=+153.972198914 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.507842 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-bcsmj" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.515086 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-cc77j"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.515392 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.526237 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.540188 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qv78n\" (UniqueName: \"kubernetes.io/projected/06e36dcf-9973-4369-8dc0-735cc42a68aa-kube-api-access-qv78n\") pod \"dns-default-9ls2l\" (UID: \"06e36dcf-9973-4369-8dc0-735cc42a68aa\") " pod="openshift-dns/dns-default-9ls2l" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.541159 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.543935 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.549060 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.560076 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pd7s\" (UniqueName: \"kubernetes.io/projected/9dbe3199-893b-4d67-a447-fa3223c321e3-kube-api-access-9pd7s\") pod \"ingress-canary-bp2vx\" (UID: \"9dbe3199-893b-4d67-a447-fa3223c321e3\") " pod="openshift-ingress-canary/ingress-canary-bp2vx" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.582939 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5jbw\" (UniqueName: \"kubernetes.io/projected/13b4eba0-42e7-448a-998c-d7aaeffc9218-kube-api-access-v5jbw\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.599578 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:24 crc kubenswrapper[4948]: E0220 08:08:24.599908 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:25.09989068 +0000 UTC m=+154.074385500 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.601898 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5kfk4" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.612653 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtrdt\" (UniqueName: \"kubernetes.io/projected/bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156-kube-api-access-qtrdt\") pod \"machine-config-server-gd69n\" (UID: \"bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156\") " pod="openshift-machine-config-operator/machine-config-server-gd69n" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.613979 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-rtnmx" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.626348 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.638314 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-9ls2l" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.652409 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.652667 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-bp2vx" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.653642 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-k8zcr"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.655327 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fl47j"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.657163 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-44l7j"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.698169 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.699836 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-gd69n" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.703790 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: E0220 08:08:24.704428 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:25.204408683 +0000 UTC m=+154.178903503 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.745847 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-xjbrz"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.777944 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-klsxg"] Feb 20 08:08:24 crc kubenswrapper[4948]: W0220 08:08:24.785486 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c070b67_c2ca_459b_a1b7_813a8833e27e.slice/crio-f24acc5a18bb988f4e5c8c67e817d3ea2636025712e6b3477a09950d4595e36b WatchSource:0}: Error finding container f24acc5a18bb988f4e5c8c67e817d3ea2636025712e6b3477a09950d4595e36b: Status 404 returned error can't find the container with id f24acc5a18bb988f4e5c8c67e817d3ea2636025712e6b3477a09950d4595e36b Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.788895 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7x85x" event={"ID":"de9b95d8-fa65-438c-bef9-2aba46044ec1","Type":"ContainerStarted","Data":"80bb30586955a2bb2425be91e482784aaca7deedc011442d5ba01e04abd2835b"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.795943 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-fhfhr"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.804647 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:24 crc kubenswrapper[4948]: E0220 08:08:24.805897 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:25.305865902 +0000 UTC m=+154.280360722 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.811312 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" event={"ID":"2c59c4e2-97ae-494a-b33a-a542266cf233","Type":"ContainerStarted","Data":"46dc1f39b430dec05ad46ae2a5d1777f0c144a8a7b1211db5a8b9a66412709c3"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.827236 4948 generic.go:334] "Generic (PLEG): container finished" podID="c7c6bf9e-0846-4391-9126-a00d68b2e627" containerID="2b2020a1e307e87d92e1c75ed8ea2540d79082ae7f10ec4114f5d6a47e553948" exitCode=0 Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.827425 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jfst2" event={"ID":"c7c6bf9e-0846-4391-9126-a00d68b2e627","Type":"ContainerDied","Data":"2b2020a1e307e87d92e1c75ed8ea2540d79082ae7f10ec4114f5d6a47e553948"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.833169 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" event={"ID":"5bc3d0cd-eacd-4b25-9acd-853e49db7b47","Type":"ContainerStarted","Data":"c3ef45d9487fa7c7fcc3b038b4c27102a6491a3966150aa1fc76eec4719774ea"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.833213 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" event={"ID":"5bc3d0cd-eacd-4b25-9acd-853e49db7b47","Type":"ContainerStarted","Data":"5a21d2c4a35b9613a1f1d569beddc744e500d407a9692ba5a38e15f50751b685"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.843475 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vtkbz" event={"ID":"fabbac48-e5db-4057-b9a5-69118c38c667","Type":"ContainerStarted","Data":"3cc6318d369798be51a8e3af59f16c5cf66e1273c8642f9c640bdcaf2fbdb905"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.843519 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vtkbz" event={"ID":"fabbac48-e5db-4057-b9a5-69118c38c667","Type":"ContainerStarted","Data":"0078d6f35d4b09153ac5f4519a90739d3b7c1a3332b872561f2b13e7ca2d48f9"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.854233 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" event={"ID":"05fe2d3b-0a53-4bbe-b0aa-be07acb159bb","Type":"ContainerStarted","Data":"e6bd43e85d4f40780c71d206d3cda54da562edaaf59358c74e9a9f95f771792a"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.873145 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs" event={"ID":"1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72","Type":"ContainerStarted","Data":"45cb715b9778f0421e1f7a41623f8cc620132945850c903c2ad1e537a17ac52f"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.873193 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs" event={"ID":"1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72","Type":"ContainerStarted","Data":"e8e9ca0e8794f3f1b33e90fba54ddb0b40a7135497207e0f2a0473343f74e227"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.876123 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-grnfc"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.891999 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-cc77j" event={"ID":"3f4bc9ec-3eda-4d56-9696-df15c14d2e4a","Type":"ContainerStarted","Data":"6a636a454976750e16ccc1469c63f538a09356f52d37fa7566be8b50b72c9709"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.907251 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl" event={"ID":"01d95eeb-e421-4fe2-a24a-44c3014f1f6e","Type":"ContainerStarted","Data":"7635ec0ec7c24c8203327d8beda70ffb895aa8078a28ac99baf419a8966811ca"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.907299 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl" event={"ID":"01d95eeb-e421-4fe2-a24a-44c3014f1f6e","Type":"ContainerStarted","Data":"fd9df30c5d63b531a8098b8313e0aead45e05e65f20b2d3a9242e2f927d8db1f"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.908349 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: E0220 08:08:24.908620 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:25.408608771 +0000 UTC m=+154.383103591 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.909125 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sjnxm"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.917275 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-x8fgj"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.942252 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nzj8d" event={"ID":"e029f4ab-6a61-4fcf-9997-b51c1cbe5675","Type":"ContainerStarted","Data":"989ca99ae7d68ec8b3478c17b0e5e440952effaa6b8e84cc390e63652c262e43"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.947177 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" event={"ID":"b1099221-b48f-4756-b133-0fed16b1e225","Type":"ContainerStarted","Data":"73250930b110d9ec15accb42a0a2a49b5ad06046d887c0bd11c8d587d39d5ab1"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.968475 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgm85" event={"ID":"a149fe37-c748-4120-9116-1da4b680d880","Type":"ContainerStarted","Data":"c6a58e0e094dca5eb92e842406fb7bb59ef92c395eadd8ac868c827c07eb9392"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.968521 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgm85" event={"ID":"a149fe37-c748-4120-9116-1da4b680d880","Type":"ContainerStarted","Data":"64d960029fc420361a87e4a17c366841af210e237e40e75ca94a40969f226b89"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.982515 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" event={"ID":"cf612779-d96d-48f1-9958-5f92ac6e3d40","Type":"ContainerStarted","Data":"d785a33073edb1bc2cec1d3eb6855e83ab9c2dd8f8168b35d03bdcffc0a1f5da"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.986965 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.993655 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-tdhsc"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.996541 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" event={"ID":"02036666-815d-4282-86f0-1b4163ea7cf9","Type":"ContainerStarted","Data":"b835c835e64d80d465d81fe9f329df642a1182ff71ea166fabeb1ad480fae83b"} Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:24.996582 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" event={"ID":"02036666-815d-4282-86f0-1b4163ea7cf9","Type":"ContainerStarted","Data":"42bd08c72bddb366038d9580429774e95c2763c7272273dc2cac5528d750116b"} Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.001811 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm"] Feb 20 08:08:25 crc kubenswrapper[4948]: W0220 08:08:25.005119 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbc13d9fe_fd58_4d3b_9278_4f3e59da5976.slice/crio-1e15ef1a019fef26569179f0c621d2a2878c8686434cf1cb9b657cedc5b91011 WatchSource:0}: Error finding container 1e15ef1a019fef26569179f0c621d2a2878c8686434cf1cb9b657cedc5b91011: Status 404 returned error can't find the container with id 1e15ef1a019fef26569179f0c621d2a2878c8686434cf1cb9b657cedc5b91011 Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.009830 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:25 crc kubenswrapper[4948]: E0220 08:08:25.010141 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:25.51011534 +0000 UTC m=+154.484610160 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.010313 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:25 crc kubenswrapper[4948]: E0220 08:08:25.011072 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:25.511061514 +0000 UTC m=+154.485556334 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.018535 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-46sg8"] Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.111220 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:25 crc kubenswrapper[4948]: E0220 08:08:25.112354 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:25.612336137 +0000 UTC m=+154.586830957 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.167563 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qgskm"] Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.201220 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk"] Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.219180 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:25 crc kubenswrapper[4948]: E0220 08:08:25.219615 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:25.719599018 +0000 UTC m=+154.694093838 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.258943 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-trh7g"] Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.320679 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:25 crc kubenswrapper[4948]: E0220 08:08:25.321215 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:25.821195599 +0000 UTC m=+154.795690419 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.430589 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:25 crc kubenswrapper[4948]: E0220 08:08:25.432355 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:25.932335285 +0000 UTC m=+154.906830095 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.552092 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:25 crc kubenswrapper[4948]: E0220 08:08:25.552585 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:26.052567503 +0000 UTC m=+155.027062323 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.648005 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl" podStartSLOduration=131.647952043 podStartE2EDuration="2m11.647952043s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:25.636908412 +0000 UTC m=+154.611403232" watchObservedRunningTime="2026-02-20 08:08:25.647952043 +0000 UTC m=+154.622446863" Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.660369 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:25 crc kubenswrapper[4948]: E0220 08:08:25.660836 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:26.160822548 +0000 UTC m=+155.135317368 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.763051 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:25 crc kubenswrapper[4948]: E0220 08:08:25.763150 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:26.263119787 +0000 UTC m=+155.237614597 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.763685 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:25 crc kubenswrapper[4948]: E0220 08:08:25.764207 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:26.264195363 +0000 UTC m=+155.238690183 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.815539 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-9ls2l"] Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.817571 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h"] Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.827498 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5kfk4"] Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.837906 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm"] Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.871512 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:25 crc kubenswrapper[4948]: E0220 08:08:25.872086 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:26.372069489 +0000 UTC m=+155.346564309 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.930728 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-bcsmj"] Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.931896 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sjv8t"] Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.959431 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rtnmx"] Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.973265 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:25 crc kubenswrapper[4948]: E0220 08:08:25.973659 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:26.47364294 +0000 UTC m=+155.448137750 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:25.997509 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs" podStartSLOduration=131.997488525 podStartE2EDuration="2m11.997488525s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:25.996112251 +0000 UTC m=+154.970607071" watchObservedRunningTime="2026-02-20 08:08:25.997488525 +0000 UTC m=+154.971983345" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.075661 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:26 crc kubenswrapper[4948]: E0220 08:08:26.076709 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:26.576689277 +0000 UTC m=+155.551184097 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.087057 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fl47j" event={"ID":"c38c7edf-ca04-4ea4-b15f-d83dea3f545e","Type":"ContainerStarted","Data":"c95f2889704f670ef930f1019a8454e6d7e06bf638a37e204f5949de0826f43c"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.087188 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fl47j" event={"ID":"c38c7edf-ca04-4ea4-b15f-d83dea3f545e","Type":"ContainerStarted","Data":"ea691caa5c1656ad51f3d8af407e41e98d60cd663139e39b82d64f3ca6e94e94"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.100519 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-bp2vx"] Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.101236 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h" event={"ID":"912a3267-581f-47b1-873c-b0e4d8dd3768","Type":"ContainerStarted","Data":"f0dd589c2408c422566835987233c145aeb400052dad0180e547063e2d0d6740"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.104213 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nzj8d" podStartSLOduration=132.104197121 podStartE2EDuration="2m12.104197121s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:26.101039094 +0000 UTC m=+155.075533914" watchObservedRunningTime="2026-02-20 08:08:26.104197121 +0000 UTC m=+155.078691961" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.134443 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs" event={"ID":"1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72","Type":"ContainerStarted","Data":"4575bc4762444482fba32a2a68c64a3d1de4d3e3bb3854c702bed016ecb3b763"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.142542 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" event={"ID":"6c070b67-c2ca-459b-a1b7-813a8833e27e","Type":"ContainerStarted","Data":"27c38be8b9e4ccb4d5b9ff6afe53e0b0eb4b257a4accae7cd8daf9b3703d5ce8"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.142597 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" event={"ID":"6c070b67-c2ca-459b-a1b7-813a8833e27e","Type":"ContainerStarted","Data":"f24acc5a18bb988f4e5c8c67e817d3ea2636025712e6b3477a09950d4595e36b"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.143754 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.164721 4948 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-qljmk container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.164787 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" podUID="6c070b67-c2ca-459b-a1b7-813a8833e27e" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.177090 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:26 crc kubenswrapper[4948]: E0220 08:08:26.177425 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:26.677411927 +0000 UTC m=+155.651906747 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.190253 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-grnfc" event={"ID":"bc13d9fe-fd58-4d3b-9278-4f3e59da5976","Type":"ContainerStarted","Data":"1e15ef1a019fef26569179f0c621d2a2878c8686434cf1cb9b657cedc5b91011"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.206344 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" event={"ID":"05fe2d3b-0a53-4bbe-b0aa-be07acb159bb","Type":"ContainerStarted","Data":"e08a5f2d9b0423afb18692d9ad335a38e239c46dd9ffe8714f61e3f2a149ddf6"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.223822 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-tdhsc" event={"ID":"40656b16-31ff-4065-b87c-5678308e9fb4","Type":"ContainerStarted","Data":"a032e77d70839df518e603667ef479b401e6f48da056e73539ba49efbcdbcc37"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.243429 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7x85x" event={"ID":"de9b95d8-fa65-438c-bef9-2aba46044ec1","Type":"ContainerStarted","Data":"e7637483275f788126b9aeb5d1bb4c5cdba980ce676e8999b4f7749cf3b9f849"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.243488 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7x85x" event={"ID":"de9b95d8-fa65-438c-bef9-2aba46044ec1","Type":"ContainerStarted","Data":"90e4321361a39876f68035a8a153661f0dc18b7c4540e6d751217c691e0141b7"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.245631 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgm85" podStartSLOduration=132.245612709 podStartE2EDuration="2m12.245612709s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:26.243234721 +0000 UTC m=+155.217729541" watchObservedRunningTime="2026-02-20 08:08:26.245612709 +0000 UTC m=+155.220107529" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.273386 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-k8zcr" event={"ID":"f81366e5-fbdb-4a0b-a91f-f603d9b8570d","Type":"ContainerStarted","Data":"61712307e24396811f0211771ba686b9bbc2eb0844719329e63b16d44d321de6"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.278576 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:26 crc kubenswrapper[4948]: E0220 08:08:26.279898 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:26.77988045 +0000 UTC m=+155.754375270 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.281112 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" event={"ID":"f77016a6-cd35-49df-b6b5-65b4858b41c9","Type":"ContainerStarted","Data":"92efa091cdddd802dc78aa2d9bf128d9d0f0ac869c640718f1c1a8d1aead053f"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.345646 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" event={"ID":"b1099221-b48f-4756-b133-0fed16b1e225","Type":"ContainerStarted","Data":"68e683b1d4b19ead7db68fed68ae56db80880a0e972aa74ee98c29b93213ec23"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.364298 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-gd69n" event={"ID":"bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156","Type":"ContainerStarted","Data":"7afe19488bbf8afd4db3d1cdf3138ac934b18ee11bcc557b188b23131293e2b0"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.381443 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:26 crc kubenswrapper[4948]: E0220 08:08:26.383243 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:26.883225324 +0000 UTC m=+155.857720144 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.404462 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-cc77j" event={"ID":"3f4bc9ec-3eda-4d56-9696-df15c14d2e4a","Type":"ContainerStarted","Data":"27e69ab691f0c40ba85f35afadd1e94a256245e171d5c78e3edbb1d1daa7626d"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.431276 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-x8fgj" event={"ID":"6f1e0caa-03ee-4025-9f2a-6941090c178b","Type":"ContainerStarted","Data":"cabbaec805df0b1917ae1dca77b867f02165355a9ddd40f61184fbc442797a01"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.466845 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" event={"ID":"cf612779-d96d-48f1-9958-5f92ac6e3d40","Type":"ContainerStarted","Data":"36f337a8a94ffe20a346a9ea0395f661eb311a188ba455aa9e3379d139dafd83"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.482765 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.482938 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" event={"ID":"5bc3d0cd-eacd-4b25-9acd-853e49db7b47","Type":"ContainerStarted","Data":"10ec9cf1938d190ec36da83cd5d3e21b002f9026662192a8c0885233df749235"} Feb 20 08:08:26 crc kubenswrapper[4948]: E0220 08:08:26.483122 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:26.983105724 +0000 UTC m=+155.957600544 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.523209 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" event={"ID":"02036666-815d-4282-86f0-1b4163ea7cf9","Type":"ContainerStarted","Data":"173d3953bfe8097f250ffa8ab48fb26722e29618fcdb6e7c03f5c33121453669"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.529858 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vtkbz" podStartSLOduration=132.52983364 podStartE2EDuration="2m12.52983364s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:26.526440416 +0000 UTC m=+155.500935236" watchObservedRunningTime="2026-02-20 08:08:26.52983364 +0000 UTC m=+155.504328460" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.566242 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-44l7j" event={"ID":"0b98dbb3-7986-475d-8028-0879d2fed2af","Type":"ContainerStarted","Data":"1f5342e62238c49aaaf9cc95280bea091a38ab7b2dccb6e71f5b3c95eea48e86"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.582781 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-trh7g" event={"ID":"13b4eba0-42e7-448a-998c-d7aaeffc9218","Type":"ContainerStarted","Data":"48fc2a7997571c36fba0d9d88c82b7bc0153d8fb89341ddef42a5a815a193a0a"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.584505 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:26 crc kubenswrapper[4948]: E0220 08:08:26.584870 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:27.084855889 +0000 UTC m=+156.059350709 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.591061 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qgskm" event={"ID":"e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf","Type":"ContainerStarted","Data":"74fbc575b905b01e85f8fd524cf1ed8beb2153e8f1ad32af0cdb433f5f045a3f"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.592641 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" event={"ID":"21db7edd-6db3-4afa-b470-0016a9c5afb9","Type":"ContainerStarted","Data":"84341e8559dd79e093cb4de6142b9a22ac6a50040f68f50623fb6ec98aa5e33a"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.596841 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.643155 4948 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-fhfhr container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.643335 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" podUID="21db7edd-6db3-4afa-b470-0016a9c5afb9" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.643683 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nzj8d" event={"ID":"e029f4ab-6a61-4fcf-9997-b51c1cbe5675","Type":"ContainerStarted","Data":"787a481b6915374903830553ffbff88aa0c720e581c4c467bfca9610de775aa3"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.686270 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" event={"ID":"df73228e-df1f-4e9c-aafe-493bb743f98a","Type":"ContainerStarted","Data":"76dafdb5407bf70db2d0cf3c70d0d7427c429ce3ab8c2b2e1a871cb56382ee9a"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.686573 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:26 crc kubenswrapper[4948]: E0220 08:08:26.686630 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:27.186614695 +0000 UTC m=+156.161109515 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.687773 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:26 crc kubenswrapper[4948]: E0220 08:08:26.689479 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:27.189461884 +0000 UTC m=+156.163956704 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.691098 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-46sg8" event={"ID":"fb7c802e-5175-4b0b-8b70-91efb1c83fa1","Type":"ContainerStarted","Data":"c4e3e88d81b8869b53780024210a6a2d4dc27ab82243c5eadb9dbdba0329ef31"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.701117 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-klsxg" event={"ID":"0b81d26a-f92a-4602-89ef-5dd9fb24a32f","Type":"ContainerStarted","Data":"47f1a06008e2047099df4b7d2d4f0a3a1cddadec6a353f1e8b9d66189bc5b84f"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.705786 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-9ls2l" event={"ID":"06e36dcf-9973-4369-8dc0-735cc42a68aa","Type":"ContainerStarted","Data":"2eea3de6acb8faeaefa1a5be0833ddd962a2c40a7011080ed01a67a8d98e108c"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.712917 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" event={"ID":"e9f46a23-4535-4d64-ae76-065435f8f762","Type":"ContainerStarted","Data":"616e47f1bb6448f32a0fe292b91640bb6d73df61cbdba5efe885de734cecb342"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.726665 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" podStartSLOduration=132.726557214 podStartE2EDuration="2m12.726557214s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:26.70560197 +0000 UTC m=+155.680096790" watchObservedRunningTime="2026-02-20 08:08:26.726557214 +0000 UTC m=+155.701052034" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.737364 4948 csr.go:261] certificate signing request csr-5cvr7 is approved, waiting to be issued Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.750269 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-xjbrz" event={"ID":"bbc3437e-8729-46f8-aee2-ad3a4679097f","Type":"ContainerStarted","Data":"c478364bdabada6a44416428ef78a5f89ef672caf5d301b485497a22b89e761e"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.751056 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-xjbrz" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.751327 4948 csr.go:257] certificate signing request csr-5cvr7 is issued Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.761462 4948 patch_prober.go:28] interesting pod/console-operator-58897d9998-xjbrz container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.761532 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-xjbrz" podUID="bbc3437e-8729-46f8-aee2-ad3a4679097f" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.763633 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7x85x" podStartSLOduration=132.763611563 podStartE2EDuration="2m12.763611563s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:26.741321206 +0000 UTC m=+155.715816026" watchObservedRunningTime="2026-02-20 08:08:26.763611563 +0000 UTC m=+155.738106373" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.766364 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-sc5pm" event={"ID":"4b3348f9-d9b9-4efb-9eb8-877f58b0d59c","Type":"ContainerStarted","Data":"ad705231484db9909f86eb5c84a06a81f6e5a0c076e43683547b25a3a95dbe85"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.770690 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" podStartSLOduration=132.770674156 podStartE2EDuration="2m12.770674156s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:26.766704559 +0000 UTC m=+155.741199369" watchObservedRunningTime="2026-02-20 08:08:26.770674156 +0000 UTC m=+155.745168976" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.790702 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:26 crc kubenswrapper[4948]: E0220 08:08:26.791994 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:27.291960088 +0000 UTC m=+156.266454898 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.792342 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm" event={"ID":"c7b67dd6-4025-45ad-98f8-2e7f8ab4b573","Type":"ContainerStarted","Data":"dc3f8115d345f257d8228dd6508393b32ec4c3016cf8750d09de9a72aa5aa83e"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.823763 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" podStartSLOduration=132.823731757 podStartE2EDuration="2m12.823731757s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:26.80630979 +0000 UTC m=+155.780804610" watchObservedRunningTime="2026-02-20 08:08:26.823731757 +0000 UTC m=+155.798226577" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.836108 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" podStartSLOduration=132.8360877 podStartE2EDuration="2m12.8360877s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:26.826639519 +0000 UTC m=+155.801134339" watchObservedRunningTime="2026-02-20 08:08:26.8360877 +0000 UTC m=+155.810582520" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.858831 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fl47j" podStartSLOduration=132.858800157 podStartE2EDuration="2m12.858800157s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:26.857438444 +0000 UTC m=+155.831933264" watchObservedRunningTime="2026-02-20 08:08:26.858800157 +0000 UTC m=+155.833294977" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.876103 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" podStartSLOduration=132.876083151 podStartE2EDuration="2m12.876083151s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:26.875495777 +0000 UTC m=+155.849990597" watchObservedRunningTime="2026-02-20 08:08:26.876083151 +0000 UTC m=+155.850577961" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.892761 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:26 crc kubenswrapper[4948]: E0220 08:08:26.893910 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:27.393895158 +0000 UTC m=+156.368389978 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.923942 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" podStartSLOduration=132.923916304 podStartE2EDuration="2m12.923916304s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:26.920587582 +0000 UTC m=+155.895082402" watchObservedRunningTime="2026-02-20 08:08:26.923916304 +0000 UTC m=+155.898411124" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.995450 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:26 crc kubenswrapper[4948]: E0220 08:08:26.995773 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:27.495736205 +0000 UTC m=+156.470231025 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.996344 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:26 crc kubenswrapper[4948]: E0220 08:08:26.996867 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:27.496850763 +0000 UTC m=+156.471345583 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.003033 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-klsxg" podStartSLOduration=133.003009834 podStartE2EDuration="2m13.003009834s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:27.00244306 +0000 UTC m=+155.976937870" watchObservedRunningTime="2026-02-20 08:08:27.003009834 +0000 UTC m=+155.977504654" Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.102505 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:27 crc kubenswrapper[4948]: E0220 08:08:27.102713 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:27.602697388 +0000 UTC m=+156.577192208 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.103171 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:27 crc kubenswrapper[4948]: E0220 08:08:27.103530 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:27.603522989 +0000 UTC m=+156.578017809 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.105875 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-sc5pm" podStartSLOduration=133.105851556 podStartE2EDuration="2m13.105851556s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:27.101649983 +0000 UTC m=+156.076144803" watchObservedRunningTime="2026-02-20 08:08:27.105851556 +0000 UTC m=+156.080346376" Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.108088 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-xjbrz" podStartSLOduration=133.108079771 podStartE2EDuration="2m13.108079771s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:27.076721631 +0000 UTC m=+156.051216451" watchObservedRunningTime="2026-02-20 08:08:27.108079771 +0000 UTC m=+156.082574591" Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.206437 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:27 crc kubenswrapper[4948]: E0220 08:08:27.207481 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:27.707449297 +0000 UTC m=+156.681944117 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.308209 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:27 crc kubenswrapper[4948]: E0220 08:08:27.308655 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:27.808638299 +0000 UTC m=+156.783133119 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:27 crc kubenswrapper[4948]: E0220 08:08:27.409128 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:27.909109683 +0000 UTC m=+156.883604503 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.409038 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.409377 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:27 crc kubenswrapper[4948]: E0220 08:08:27.409724 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:27.909713558 +0000 UTC m=+156.884208378 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.463058 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.463454 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.463482 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.512855 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:27 crc kubenswrapper[4948]: E0220 08:08:27.513866 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:28.013846831 +0000 UTC m=+156.988341651 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.615102 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:27 crc kubenswrapper[4948]: E0220 08:08:27.615489 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:28.115475764 +0000 UTC m=+157.089970584 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.717203 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:27 crc kubenswrapper[4948]: E0220 08:08:27.717554 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:28.217523356 +0000 UTC m=+157.192018176 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.719192 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:27 crc kubenswrapper[4948]: E0220 08:08:27.719687 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:28.219670649 +0000 UTC m=+157.194165469 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.757062 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-20 08:03:26 +0000 UTC, rotation deadline is 2026-12-18 03:05:35.800339458 +0000 UTC Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.757412 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7218h57m8.042930834s for next certificate rotation Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.801827 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-tdhsc" event={"ID":"40656b16-31ff-4065-b87c-5678308e9fb4","Type":"ContainerStarted","Data":"8010bfb6dfead3b5a6d77b49e41e950061de92d8c07b3666d38680482a3844af"} Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.808908 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-bcsmj" event={"ID":"000483d2-8b2d-4403-9652-d4d86d55f7b5","Type":"ContainerStarted","Data":"8142fe9f926bd6c4b6cdd3b8e3dc4f48366e5d712648e9b3e5e5813d7095a93c"} Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.808956 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-bcsmj" event={"ID":"000483d2-8b2d-4403-9652-d4d86d55f7b5","Type":"ContainerStarted","Data":"e610ee04d021213df53cb4b947552b128f5bd097bb682e7b6347c12abee71f0f"} Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.820784 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:27 crc kubenswrapper[4948]: E0220 08:08:27.822228 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:28.322203584 +0000 UTC m=+157.296698394 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.825024 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-tdhsc" podStartSLOduration=133.824998792 podStartE2EDuration="2m13.824998792s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:27.824417948 +0000 UTC m=+156.798912768" watchObservedRunningTime="2026-02-20 08:08:27.824998792 +0000 UTC m=+156.799493612" Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.830052 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" event={"ID":"df73228e-df1f-4e9c-aafe-493bb743f98a","Type":"ContainerStarted","Data":"f0e4fad62991d25dea391bf5b9e653af0d194dd88ba8a814b4a27d229ce2ef30"} Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.831153 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.843718 4948 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-sjnxm container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.33:6443/healthz\": dial tcp 10.217.0.33:6443: connect: connection refused" start-of-body= Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.844060 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" podUID="df73228e-df1f-4e9c-aafe-493bb743f98a" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.33:6443/healthz\": dial tcp 10.217.0.33:6443: connect: connection refused" Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.853427 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qgskm" event={"ID":"e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf","Type":"ContainerStarted","Data":"158588997f0116bf9c80751e42eec3bc450832a51470d5e36f65ed2c94dbeb58"} Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.866051 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" podStartSLOduration=133.866019898 podStartE2EDuration="2m13.866019898s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:27.858192406 +0000 UTC m=+156.832687226" watchObservedRunningTime="2026-02-20 08:08:27.866019898 +0000 UTC m=+156.840514718" Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.874206 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-46sg8" event={"ID":"fb7c802e-5175-4b0b-8b70-91efb1c83fa1","Type":"ContainerStarted","Data":"c3dce3a4ba3c8fc0032fff443af8742149cd5cdf42877e90386f11470aa73932"} Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.874258 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-46sg8" event={"ID":"fb7c802e-5175-4b0b-8b70-91efb1c83fa1","Type":"ContainerStarted","Data":"25c94662e3decbb3b4e2aaba9a0cb2773839779bffa1f9f9a23009b452d4dfcd"} Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.874941 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-46sg8" Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.910481 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-bp2vx" event={"ID":"9dbe3199-893b-4d67-a447-fa3223c321e3","Type":"ContainerStarted","Data":"92a8b354f092998d33eb260a3546688c2937af73a1b2ae389b0bea11d8f50794"} Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.914610 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-bp2vx" event={"ID":"9dbe3199-893b-4d67-a447-fa3223c321e3","Type":"ContainerStarted","Data":"63e9556a2a22b7fd0f1d180db3021b4e14c73e783441d8d43c988d3dd1c01136"} Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.922050 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:27 crc kubenswrapper[4948]: E0220 08:08:27.926443 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:28.426415838 +0000 UTC m=+157.400910648 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.944991 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qgskm" podStartSLOduration=133.944947363 podStartE2EDuration="2m13.944947363s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:27.88813299 +0000 UTC m=+156.862627810" watchObservedRunningTime="2026-02-20 08:08:27.944947363 +0000 UTC m=+156.919442183" Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.968377 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" event={"ID":"b1099221-b48f-4756-b133-0fed16b1e225","Type":"ContainerStarted","Data":"f602ad97a5fc08e4d185a5a9671744a590f1170b60b5ca4802d6f917630253cf"} Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.978230 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-46sg8" podStartSLOduration=133.978210818 podStartE2EDuration="2m13.978210818s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:27.947505005 +0000 UTC m=+156.921999825" watchObservedRunningTime="2026-02-20 08:08:27.978210818 +0000 UTC m=+156.952705638" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.012807 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-bp2vx" podStartSLOduration=7.012784476 podStartE2EDuration="7.012784476s" podCreationTimestamp="2026-02-20 08:08:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:27.978958287 +0000 UTC m=+156.953453117" watchObservedRunningTime="2026-02-20 08:08:28.012784476 +0000 UTC m=+156.987279296" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.013912 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" podStartSLOduration=134.013907224 podStartE2EDuration="2m14.013907224s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:28.01088422 +0000 UTC m=+156.985379040" watchObservedRunningTime="2026-02-20 08:08:28.013907224 +0000 UTC m=+156.988402034" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.014572 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-sc5pm" event={"ID":"4b3348f9-d9b9-4efb-9eb8-877f58b0d59c","Type":"ContainerStarted","Data":"c749eafa49326acbaefed5b7bdb334d8f07a133966a45121fbd7300edd8f537b"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.031473 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.037302 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" event={"ID":"f77016a6-cd35-49df-b6b5-65b4858b41c9","Type":"ContainerStarted","Data":"0520ff9e3a8f12a5228a352b59373978c12d5f10bb51923b12eab15acf564086"} Feb 20 08:08:28 crc kubenswrapper[4948]: E0220 08:08:28.038399 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:28.538379204 +0000 UTC m=+157.512874024 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.059290 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-xjbrz" event={"ID":"bbc3437e-8729-46f8-aee2-ad3a4679097f","Type":"ContainerStarted","Data":"e0124670c1008f3fd7929e53fe9680fae097c86e551ab7d64a8f6733d6449e84"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.070025 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" podStartSLOduration=134.07000051 podStartE2EDuration="2m14.07000051s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:28.068913163 +0000 UTC m=+157.043407983" watchObservedRunningTime="2026-02-20 08:08:28.07000051 +0000 UTC m=+157.044495330" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.081345 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-klsxg" event={"ID":"0b81d26a-f92a-4602-89ef-5dd9fb24a32f","Type":"ContainerStarted","Data":"b029d80c1e577342eaa2888ffe51427b6da3e30eb7f1e2bafc7ab86e988b66c1"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.092683 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm" event={"ID":"c7b67dd6-4025-45ad-98f8-2e7f8ab4b573","Type":"ContainerStarted","Data":"f08b11d190142a1c0ac494119229274702a7dd496184ab826a44da91c5aeb639"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.093350 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.103376 4948 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-6zdkm container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.103438 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm" podUID="c7b67dd6-4025-45ad-98f8-2e7f8ab4b573" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.115271 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-rtnmx" event={"ID":"308389f1-6962-45be-87b0-1b061caa2c50","Type":"ContainerStarted","Data":"0485996e9a881ae3c35cf440e5293526ff386a26f8480ffec43777bc9f6cd2cc"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.115475 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-rtnmx" event={"ID":"308389f1-6962-45be-87b0-1b061caa2c50","Type":"ContainerStarted","Data":"406d80af236a11af7cfd4a41db7671d2067912bafd6e7cebcc426a1590e948e3"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.135448 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:28 crc kubenswrapper[4948]: E0220 08:08:28.136406 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:28.636384217 +0000 UTC m=+157.610879107 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.149624 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" event={"ID":"95a270e5-7a08-4238-a65d-96986f2d2c39","Type":"ContainerStarted","Data":"15687151f4ed59c06cf2071d3204504981cf4a0c6cdb40d25a5dbd083313ed65"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.149946 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" event={"ID":"95a270e5-7a08-4238-a65d-96986f2d2c39","Type":"ContainerStarted","Data":"936301da215b82b117fb711a926cf169f025d16470e0871be5907e8efa5aec8e"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.151736 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.158487 4948 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-sjv8t container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.159013 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" podUID="95a270e5-7a08-4238-a65d-96986f2d2c39" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.162510 4948 generic.go:334] "Generic (PLEG): container finished" podID="2c59c4e2-97ae-494a-b33a-a542266cf233" containerID="b54927298d346e7bc2198cc485b8d321498aea49b86a1ada6f9600eb7b7e5912" exitCode=0 Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.162708 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" event={"ID":"2c59c4e2-97ae-494a-b33a-a542266cf233","Type":"ContainerDied","Data":"b54927298d346e7bc2198cc485b8d321498aea49b86a1ada6f9600eb7b7e5912"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.177121 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-rtnmx" podStartSLOduration=133.177093826 podStartE2EDuration="2m13.177093826s" podCreationTimestamp="2026-02-20 08:06:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:28.176518342 +0000 UTC m=+157.151013162" watchObservedRunningTime="2026-02-20 08:08:28.177093826 +0000 UTC m=+157.151588646" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.178552 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm" podStartSLOduration=134.178542891 podStartE2EDuration="2m14.178542891s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:28.128936175 +0000 UTC m=+157.103430995" watchObservedRunningTime="2026-02-20 08:08:28.178542891 +0000 UTC m=+157.153037711" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.189858 4948 generic.go:334] "Generic (PLEG): container finished" podID="0b98dbb3-7986-475d-8028-0879d2fed2af" containerID="ddd4a11d3470308b5ca2eb6c345ef2617f7ff6bae43e4451822a612395bb815e" exitCode=0 Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.189930 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-44l7j" event={"ID":"0b98dbb3-7986-475d-8028-0879d2fed2af","Type":"ContainerDied","Data":"ddd4a11d3470308b5ca2eb6c345ef2617f7ff6bae43e4451822a612395bb815e"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.219230 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" event={"ID":"21db7edd-6db3-4afa-b470-0016a9c5afb9","Type":"ContainerStarted","Data":"aa75719650689fb107484882a72b15a095aba1554265c006218efdf71c2e2e3e"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.239319 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:28 crc kubenswrapper[4948]: E0220 08:08:28.239547 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:28.739513937 +0000 UTC m=+157.714008757 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.241243 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:28 crc kubenswrapper[4948]: E0220 08:08:28.257582 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:28.757566079 +0000 UTC m=+157.732060899 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.280789 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" podStartSLOduration=134.280771718 podStartE2EDuration="2m14.280771718s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:28.206188689 +0000 UTC m=+157.180683509" watchObservedRunningTime="2026-02-20 08:08:28.280771718 +0000 UTC m=+157.255266538" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.292640 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.292766 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" event={"ID":"e9f46a23-4535-4d64-ae76-065435f8f762","Type":"ContainerStarted","Data":"437f16fdb8670c6f1795395ab53f0a3d4b705057dc4b0f5bd948256585c1053e"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.292841 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" event={"ID":"e9f46a23-4535-4d64-ae76-065435f8f762","Type":"ContainerStarted","Data":"2afb989988045404096c867b49d32b1ae96c6879a5066aa9d967cf42bdd1216e"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.331309 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-x8fgj" event={"ID":"6f1e0caa-03ee-4025-9f2a-6941090c178b","Type":"ContainerStarted","Data":"62d64849daf34c87ca026bc1f7a1b40fff2ddf3be9856fd2f487a5a16d63e71b"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.342452 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:28 crc kubenswrapper[4948]: E0220 08:08:28.342778 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:28.842732628 +0000 UTC m=+157.817227448 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.343251 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:28 crc kubenswrapper[4948]: E0220 08:08:28.343804 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:28.843796044 +0000 UTC m=+157.818290864 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.392499 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jfst2" event={"ID":"c7c6bf9e-0846-4391-9126-a00d68b2e627","Type":"ContainerStarted","Data":"a33cdea76478f468d14acad0119d166426114e9269318de6b696bc966c40872c"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.392552 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jfst2" event={"ID":"c7c6bf9e-0846-4391-9126-a00d68b2e627","Type":"ContainerStarted","Data":"587dd77e32582f1cad315db93f2588f0ff6e4dc9474fde9a88701bb7760bd9cd"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.409735 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" podStartSLOduration=134.409718521 podStartE2EDuration="2m14.409718521s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:28.371534754 +0000 UTC m=+157.346029564" watchObservedRunningTime="2026-02-20 08:08:28.409718521 +0000 UTC m=+157.384213341" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.413758 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-grnfc" event={"ID":"bc13d9fe-fd58-4d3b-9278-4f3e59da5976","Type":"ContainerStarted","Data":"3d4a3a3e86ff6555077f70891860bf39b098a33875c10bf8c2fe774c9bc00ee1"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.414575 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-grnfc" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.422253 4948 patch_prober.go:28] interesting pod/downloads-7954f5f757-grnfc container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.422334 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-grnfc" podUID="bc13d9fe-fd58-4d3b-9278-4f3e59da5976" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.440192 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" event={"ID":"a102d1fa-69a5-4826-ac40-e07ab4087558","Type":"ContainerStarted","Data":"04b654d2c614fb8a9d9b11bb1558405cff6b9ad11400bafa7c2f65d293f00646"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.440241 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" event={"ID":"a102d1fa-69a5-4826-ac40-e07ab4087558","Type":"ContainerStarted","Data":"15a04fb4755610f48f1996977d16d9994621f61b4a6425ddf01ea0f7379ae164"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.441013 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.441898 4948 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-dkdgm container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": dial tcp 10.217.0.34:5443: connect: connection refused" start-of-body= Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.441936 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" podUID="a102d1fa-69a5-4826-ac40-e07ab4087558" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": dial tcp 10.217.0.34:5443: connect: connection refused" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.444185 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:28 crc kubenswrapper[4948]: E0220 08:08:28.445401 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:28.945385405 +0000 UTC m=+157.919880235 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.460155 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5kfk4" event={"ID":"99fafd36-91d7-45b8-a321-59307d79e5ce","Type":"ContainerStarted","Data":"d2538b6448874d66303a896eee725d42cff670ea7e0b1b3eaf8c102491876a77"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.460640 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5kfk4" event={"ID":"99fafd36-91d7-45b8-a321-59307d79e5ce","Type":"ContainerStarted","Data":"daae403b300c527e65888b7699ef958c2cc74f2d75fdb04a1659c72e93a624d5"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.472202 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 08:08:28 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Feb 20 08:08:28 crc kubenswrapper[4948]: [+]process-running ok Feb 20 08:08:28 crc kubenswrapper[4948]: healthz check failed Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.472261 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.511308 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-k8zcr" event={"ID":"f81366e5-fbdb-4a0b-a91f-f603d9b8570d","Type":"ContainerStarted","Data":"3745ecb9e997c9c1a842fb71042af1df4c9828447494e128f09e0903f609bd78"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.546787 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.554173 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h" event={"ID":"912a3267-581f-47b1-873c-b0e4d8dd3768","Type":"ContainerStarted","Data":"7bcd605d1ba730936d407eda494a4e9c0c2e556a56b4acbadefc6d4d5fbdf721"} Feb 20 08:08:28 crc kubenswrapper[4948]: E0220 08:08:28.555158 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:29.055142087 +0000 UTC m=+158.029636907 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.555150 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.563160 4948 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-dgw2h container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.563231 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h" podUID="912a3267-581f-47b1-873c-b0e4d8dd3768" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.587345 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-x8fgj" podStartSLOduration=134.587329316 podStartE2EDuration="2m14.587329316s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:28.584759253 +0000 UTC m=+157.559254063" watchObservedRunningTime="2026-02-20 08:08:28.587329316 +0000 UTC m=+157.561824136" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.602292 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-gd69n" event={"ID":"bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156","Type":"ContainerStarted","Data":"5ccd94cad6ceb1d22526a784585976924a9fad67a354bd99cc01a2852b32de5c"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.639797 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-cc77j" event={"ID":"3f4bc9ec-3eda-4d56-9696-df15c14d2e4a","Type":"ContainerStarted","Data":"34cf850eb5acf1ade6721905d738cb714924f6cf6b9e924fef7ceb56f03460e0"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.651345 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.651897 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:28 crc kubenswrapper[4948]: E0220 08:08:28.653589 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:29.153564631 +0000 UTC m=+158.128059451 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.764771 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:28 crc kubenswrapper[4948]: E0220 08:08:28.766943 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:29.266928181 +0000 UTC m=+158.241423001 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.852321 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" podStartSLOduration=134.852295714 podStartE2EDuration="2m14.852295714s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:28.762956723 +0000 UTC m=+157.737451543" watchObservedRunningTime="2026-02-20 08:08:28.852295714 +0000 UTC m=+157.826790534" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.853210 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5kfk4" podStartSLOduration=134.853203137 podStartE2EDuration="2m14.853203137s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:28.837616914 +0000 UTC m=+157.812111734" watchObservedRunningTime="2026-02-20 08:08:28.853203137 +0000 UTC m=+157.827697957" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.885012 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:28 crc kubenswrapper[4948]: E0220 08:08:28.885326 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:29.385309564 +0000 UTC m=+158.359804384 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.979403 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-xjbrz" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.988771 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-grnfc" podStartSLOduration=134.98873926 podStartE2EDuration="2m14.98873926s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:28.898047376 +0000 UTC m=+157.872542196" watchObservedRunningTime="2026-02-20 08:08:28.98873926 +0000 UTC m=+157.963234080" Feb 20 08:08:28 crc kubenswrapper[4948]: E0220 08:08:28.989144 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:29.4891312 +0000 UTC m=+158.463626020 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.990313 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-jfst2" podStartSLOduration=134.990307199 podStartE2EDuration="2m14.990307199s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:28.988511215 +0000 UTC m=+157.963006035" watchObservedRunningTime="2026-02-20 08:08:28.990307199 +0000 UTC m=+157.964802009" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.988828 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.092717 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:29 crc kubenswrapper[4948]: E0220 08:08:29.093111 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:29.59309418 +0000 UTC m=+158.567589000 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.140634 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-gd69n" podStartSLOduration=8.140612855 podStartE2EDuration="8.140612855s" podCreationTimestamp="2026-02-20 08:08:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:29.074254308 +0000 UTC m=+158.048749128" watchObservedRunningTime="2026-02-20 08:08:29.140612855 +0000 UTC m=+158.115107675" Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.141510 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-k8zcr" podStartSLOduration=135.141505527 podStartE2EDuration="2m15.141505527s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:29.139209851 +0000 UTC m=+158.113704671" watchObservedRunningTime="2026-02-20 08:08:29.141505527 +0000 UTC m=+158.116000347" Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.194636 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:29 crc kubenswrapper[4948]: E0220 08:08:29.195202 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:29.695176753 +0000 UTC m=+158.669671763 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.226048 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-cc77j" podStartSLOduration=135.226020559 podStartE2EDuration="2m15.226020559s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:29.225739263 +0000 UTC m=+158.200234083" watchObservedRunningTime="2026-02-20 08:08:29.226020559 +0000 UTC m=+158.200515369" Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.295141 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:29 crc kubenswrapper[4948]: E0220 08:08:29.295448 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:29.795408601 +0000 UTC m=+158.769903421 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.295523 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:29 crc kubenswrapper[4948]: E0220 08:08:29.296175 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:29.796155189 +0000 UTC m=+158.770650009 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.327086 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h" podStartSLOduration=135.327064847 podStartE2EDuration="2m15.327064847s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:29.322742411 +0000 UTC m=+158.297237231" watchObservedRunningTime="2026-02-20 08:08:29.327064847 +0000 UTC m=+158.301559667" Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.396819 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:29 crc kubenswrapper[4948]: E0220 08:08:29.397063 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:29.897032553 +0000 UTC m=+158.871527373 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.397121 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:29 crc kubenswrapper[4948]: E0220 08:08:29.397463 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:29.897450643 +0000 UTC m=+158.871945463 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.466736 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 08:08:29 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Feb 20 08:08:29 crc kubenswrapper[4948]: [+]process-running ok Feb 20 08:08:29 crc kubenswrapper[4948]: healthz check failed Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.467349 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.498803 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:29 crc kubenswrapper[4948]: E0220 08:08:29.499048 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:29.999010174 +0000 UTC m=+158.973504994 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.499106 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:29 crc kubenswrapper[4948]: E0220 08:08:29.499596 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:29.999587168 +0000 UTC m=+158.974081988 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.599771 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:29 crc kubenswrapper[4948]: E0220 08:08:29.599911 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:30.099890318 +0000 UTC m=+159.074385138 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.600145 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:29 crc kubenswrapper[4948]: E0220 08:08:29.600446 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:30.100437971 +0000 UTC m=+159.074932791 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.665075 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-9ls2l" event={"ID":"06e36dcf-9973-4369-8dc0-735cc42a68aa","Type":"ContainerStarted","Data":"a3bbdaa9b0b3861669b1c132abe048f8076108b0936f02380e364cfdff3bc95d"} Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.665127 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-9ls2l" event={"ID":"06e36dcf-9973-4369-8dc0-735cc42a68aa","Type":"ContainerStarted","Data":"2ed05c7268732ababe3198649757956375bf224a81ac92918b4581897f5a57c1"} Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.665938 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-9ls2l" Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.671353 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" event={"ID":"2c59c4e2-97ae-494a-b33a-a542266cf233","Type":"ContainerStarted","Data":"4e297c40eb98b8058e01dae8aa3d6706d4903ebaed0c20f50c84bc44a9a72a3e"} Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.689748 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-44l7j" event={"ID":"0b98dbb3-7986-475d-8028-0879d2fed2af","Type":"ContainerStarted","Data":"7ef4f62f031aa5805a790ed6e9c2a2f08a724dd6e1a4e587caa0b73ae425de22"} Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.689794 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-44l7j" Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.695314 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-9ls2l" podStartSLOduration=8.695297658 podStartE2EDuration="8.695297658s" podCreationTimestamp="2026-02-20 08:08:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:29.69253064 +0000 UTC m=+158.667025450" watchObservedRunningTime="2026-02-20 08:08:29.695297658 +0000 UTC m=+158.669792478" Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.701047 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:29 crc kubenswrapper[4948]: E0220 08:08:29.701496 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:30.201479679 +0000 UTC m=+159.175974499 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.705190 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-trh7g" event={"ID":"13b4eba0-42e7-448a-998c-d7aaeffc9218","Type":"ContainerStarted","Data":"e6c7793a919930d4f4d795af5b8190d1891a3304b2e5950965a99e5b1db6a306"} Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.705237 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-trh7g" event={"ID":"13b4eba0-42e7-448a-998c-d7aaeffc9218","Type":"ContainerStarted","Data":"9a6fa9fe46f8ef81c7a6d3bc9620483e487ee8d5fa786777059dcee625c46d19"} Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.765138 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-44l7j" podStartSLOduration=135.76511369 podStartE2EDuration="2m15.76511369s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:29.763552172 +0000 UTC m=+158.738046992" watchObservedRunningTime="2026-02-20 08:08:29.76511369 +0000 UTC m=+158.739608510" Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.766113 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-bcsmj" event={"ID":"000483d2-8b2d-4403-9652-d4d86d55f7b5","Type":"ContainerStarted","Data":"5cc9da3f8414ed4e7f3bdc4ca498f288ed0b5722714cb01a26502c4c4a6e3f99"} Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.783916 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-x8fgj" event={"ID":"6f1e0caa-03ee-4025-9f2a-6941090c178b","Type":"ContainerStarted","Data":"47920ed6007ba3c92b551410a6ae9dc8f5f9996533e762ce51d250d6e5f0b84f"} Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.789456 4948 patch_prober.go:28] interesting pod/downloads-7954f5f757-grnfc container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.789564 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-grnfc" podUID="bc13d9fe-fd58-4d3b-9278-4f3e59da5976" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.793239 4948 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-sjv8t container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.793294 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" podUID="95a270e5-7a08-4238-a65d-96986f2d2c39" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.800747 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h" Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.801122 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.801987 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:29 crc kubenswrapper[4948]: E0220 08:08:29.803256 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:30.303231705 +0000 UTC m=+159.277726695 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.844371 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm" Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.895260 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" podStartSLOduration=135.895244301 podStartE2EDuration="2m15.895244301s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:29.894681707 +0000 UTC m=+158.869176527" watchObservedRunningTime="2026-02-20 08:08:29.895244301 +0000 UTC m=+158.869739121" Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.905495 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:29 crc kubenswrapper[4948]: E0220 08:08:29.905782 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:30.405761539 +0000 UTC m=+159.380256359 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.906397 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:29 crc kubenswrapper[4948]: E0220 08:08:29.934830 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:30.434802741 +0000 UTC m=+159.409297561 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.011800 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:30 crc kubenswrapper[4948]: E0220 08:08:30.012383 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:30.512344643 +0000 UTC m=+159.486839463 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.113413 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:30 crc kubenswrapper[4948]: E0220 08:08:30.113802 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:30.613787631 +0000 UTC m=+159.588282441 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.186524 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-bcsmj" podStartSLOduration=136.186502674 podStartE2EDuration="2m16.186502674s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:30.182914736 +0000 UTC m=+159.157409556" watchObservedRunningTime="2026-02-20 08:08:30.186502674 +0000 UTC m=+159.160997494" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.214584 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:30 crc kubenswrapper[4948]: E0220 08:08:30.214760 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:30.714733776 +0000 UTC m=+159.689228596 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.214865 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:30 crc kubenswrapper[4948]: E0220 08:08:30.215228 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:30.715213078 +0000 UTC m=+159.689707898 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.315946 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:30 crc kubenswrapper[4948]: E0220 08:08:30.316208 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:30.816168164 +0000 UTC m=+159.790662984 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.316892 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:30 crc kubenswrapper[4948]: E0220 08:08:30.317404 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:30.817387474 +0000 UTC m=+159.791882294 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.419415 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:30 crc kubenswrapper[4948]: E0220 08:08:30.420426 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:30.92038598 +0000 UTC m=+159.894880800 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.420745 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:30 crc kubenswrapper[4948]: E0220 08:08:30.421552 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:30.921514937 +0000 UTC m=+159.896009757 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.442715 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nrhl6"] Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.444290 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nrhl6" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.446728 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.460657 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nrhl6"] Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.468227 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 08:08:30 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Feb 20 08:08:30 crc kubenswrapper[4948]: [+]process-running ok Feb 20 08:08:30 crc kubenswrapper[4948]: healthz check failed Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.468300 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.524682 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.525048 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f75a50cf-7dae-420b-a00d-671ede3cb6f7-catalog-content\") pod \"certified-operators-nrhl6\" (UID: \"f75a50cf-7dae-420b-a00d-671ede3cb6f7\") " pod="openshift-marketplace/certified-operators-nrhl6" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.525129 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nc669\" (UniqueName: \"kubernetes.io/projected/f75a50cf-7dae-420b-a00d-671ede3cb6f7-kube-api-access-nc669\") pod \"certified-operators-nrhl6\" (UID: \"f75a50cf-7dae-420b-a00d-671ede3cb6f7\") " pod="openshift-marketplace/certified-operators-nrhl6" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.525152 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f75a50cf-7dae-420b-a00d-671ede3cb6f7-utilities\") pod \"certified-operators-nrhl6\" (UID: \"f75a50cf-7dae-420b-a00d-671ede3cb6f7\") " pod="openshift-marketplace/certified-operators-nrhl6" Feb 20 08:08:30 crc kubenswrapper[4948]: E0220 08:08:30.525246 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:31.02520461 +0000 UTC m=+159.999699430 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.627324 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nc669\" (UniqueName: \"kubernetes.io/projected/f75a50cf-7dae-420b-a00d-671ede3cb6f7-kube-api-access-nc669\") pod \"certified-operators-nrhl6\" (UID: \"f75a50cf-7dae-420b-a00d-671ede3cb6f7\") " pod="openshift-marketplace/certified-operators-nrhl6" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.627390 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f75a50cf-7dae-420b-a00d-671ede3cb6f7-utilities\") pod \"certified-operators-nrhl6\" (UID: \"f75a50cf-7dae-420b-a00d-671ede3cb6f7\") " pod="openshift-marketplace/certified-operators-nrhl6" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.627475 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f75a50cf-7dae-420b-a00d-671ede3cb6f7-catalog-content\") pod \"certified-operators-nrhl6\" (UID: \"f75a50cf-7dae-420b-a00d-671ede3cb6f7\") " pod="openshift-marketplace/certified-operators-nrhl6" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.627512 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:30 crc kubenswrapper[4948]: E0220 08:08:30.627925 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:31.127908619 +0000 UTC m=+160.102403439 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.628240 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f75a50cf-7dae-420b-a00d-671ede3cb6f7-utilities\") pod \"certified-operators-nrhl6\" (UID: \"f75a50cf-7dae-420b-a00d-671ede3cb6f7\") " pod="openshift-marketplace/certified-operators-nrhl6" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.628344 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f75a50cf-7dae-420b-a00d-671ede3cb6f7-catalog-content\") pod \"certified-operators-nrhl6\" (UID: \"f75a50cf-7dae-420b-a00d-671ede3cb6f7\") " pod="openshift-marketplace/certified-operators-nrhl6" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.687850 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nc669\" (UniqueName: \"kubernetes.io/projected/f75a50cf-7dae-420b-a00d-671ede3cb6f7-kube-api-access-nc669\") pod \"certified-operators-nrhl6\" (UID: \"f75a50cf-7dae-420b-a00d-671ede3cb6f7\") " pod="openshift-marketplace/certified-operators-nrhl6" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.689868 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-sbf5n"] Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.690966 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sbf5n" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.719264 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sbf5n"] Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.729022 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:30 crc kubenswrapper[4948]: E0220 08:08:30.729233 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:31.229205583 +0000 UTC m=+160.203700413 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.729343 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:30 crc kubenswrapper[4948]: E0220 08:08:30.729711 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:31.229704055 +0000 UTC m=+160.204198875 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.771129 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nrhl6" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.792257 4948 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-dkdgm container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.792658 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" podUID="a102d1fa-69a5-4826-ac40-e07ab4087558" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.809285 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-trh7g" event={"ID":"13b4eba0-42e7-448a-998c-d7aaeffc9218","Type":"ContainerStarted","Data":"8e92df1cce5fa37a86f5cbf1a1b9867cf0118f67442aabc21aee1d5d906263df"} Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.810149 4948 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-sjv8t container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.810214 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" podUID="95a270e5-7a08-4238-a65d-96986f2d2c39" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.814122 4948 patch_prober.go:28] interesting pod/downloads-7954f5f757-grnfc container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.814190 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-grnfc" podUID="bc13d9fe-fd58-4d3b-9278-4f3e59da5976" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.832794 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.833052 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/000032f5-8df9-4879-8a8c-7ca06deb0862-utilities\") pod \"certified-operators-sbf5n\" (UID: \"000032f5-8df9-4879-8a8c-7ca06deb0862\") " pod="openshift-marketplace/certified-operators-sbf5n" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.833123 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqxvg\" (UniqueName: \"kubernetes.io/projected/000032f5-8df9-4879-8a8c-7ca06deb0862-kube-api-access-dqxvg\") pod \"certified-operators-sbf5n\" (UID: \"000032f5-8df9-4879-8a8c-7ca06deb0862\") " pod="openshift-marketplace/certified-operators-sbf5n" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.833155 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/000032f5-8df9-4879-8a8c-7ca06deb0862-catalog-content\") pod \"certified-operators-sbf5n\" (UID: \"000032f5-8df9-4879-8a8c-7ca06deb0862\") " pod="openshift-marketplace/certified-operators-sbf5n" Feb 20 08:08:30 crc kubenswrapper[4948]: E0220 08:08:30.833217 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:31.333185373 +0000 UTC m=+160.307680373 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.833276 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:30 crc kubenswrapper[4948]: E0220 08:08:30.833757 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:31.333748297 +0000 UTC m=+160.308243127 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.899145 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.935058 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:30 crc kubenswrapper[4948]: E0220 08:08:30.935685 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:31.435645146 +0000 UTC m=+160.410139966 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.937946 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/000032f5-8df9-4879-8a8c-7ca06deb0862-utilities\") pod \"certified-operators-sbf5n\" (UID: \"000032f5-8df9-4879-8a8c-7ca06deb0862\") " pod="openshift-marketplace/certified-operators-sbf5n" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.938151 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqxvg\" (UniqueName: \"kubernetes.io/projected/000032f5-8df9-4879-8a8c-7ca06deb0862-kube-api-access-dqxvg\") pod \"certified-operators-sbf5n\" (UID: \"000032f5-8df9-4879-8a8c-7ca06deb0862\") " pod="openshift-marketplace/certified-operators-sbf5n" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.938248 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/000032f5-8df9-4879-8a8c-7ca06deb0862-catalog-content\") pod \"certified-operators-sbf5n\" (UID: \"000032f5-8df9-4879-8a8c-7ca06deb0862\") " pod="openshift-marketplace/certified-operators-sbf5n" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.940081 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/000032f5-8df9-4879-8a8c-7ca06deb0862-catalog-content\") pod \"certified-operators-sbf5n\" (UID: \"000032f5-8df9-4879-8a8c-7ca06deb0862\") " pod="openshift-marketplace/certified-operators-sbf5n" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.941664 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/000032f5-8df9-4879-8a8c-7ca06deb0862-utilities\") pod \"certified-operators-sbf5n\" (UID: \"000032f5-8df9-4879-8a8c-7ca06deb0862\") " pod="openshift-marketplace/certified-operators-sbf5n" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.956304 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.041066 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.041525 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqxvg\" (UniqueName: \"kubernetes.io/projected/000032f5-8df9-4879-8a8c-7ca06deb0862-kube-api-access-dqxvg\") pod \"certified-operators-sbf5n\" (UID: \"000032f5-8df9-4879-8a8c-7ca06deb0862\") " pod="openshift-marketplace/certified-operators-sbf5n" Feb 20 08:08:31 crc kubenswrapper[4948]: E0220 08:08:31.041838 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:31.54182311 +0000 UTC m=+160.516317930 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.083223 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-n8d7m"] Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.087013 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n8d7m" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.090959 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.091693 4948 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.121895 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n8d7m"] Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.144732 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:31 crc kubenswrapper[4948]: E0220 08:08:31.145303 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:31.645263637 +0000 UTC m=+160.619758467 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.145785 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-utilities\") pod \"community-operators-n8d7m\" (UID: \"00f4b4c1-af36-4ccd-a3d1-dd6df1186338\") " pod="openshift-marketplace/community-operators-n8d7m" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.145961 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.146141 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxq9k\" (UniqueName: \"kubernetes.io/projected/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-kube-api-access-kxq9k\") pod \"community-operators-n8d7m\" (UID: \"00f4b4c1-af36-4ccd-a3d1-dd6df1186338\") " pod="openshift-marketplace/community-operators-n8d7m" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.146235 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-catalog-content\") pod \"community-operators-n8d7m\" (UID: \"00f4b4c1-af36-4ccd-a3d1-dd6df1186338\") " pod="openshift-marketplace/community-operators-n8d7m" Feb 20 08:08:31 crc kubenswrapper[4948]: E0220 08:08:31.146744 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:31.646725242 +0000 UTC m=+160.621220272 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.248680 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.249266 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxq9k\" (UniqueName: \"kubernetes.io/projected/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-kube-api-access-kxq9k\") pod \"community-operators-n8d7m\" (UID: \"00f4b4c1-af36-4ccd-a3d1-dd6df1186338\") " pod="openshift-marketplace/community-operators-n8d7m" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.249307 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-catalog-content\") pod \"community-operators-n8d7m\" (UID: \"00f4b4c1-af36-4ccd-a3d1-dd6df1186338\") " pod="openshift-marketplace/community-operators-n8d7m" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.249340 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-utilities\") pod \"community-operators-n8d7m\" (UID: \"00f4b4c1-af36-4ccd-a3d1-dd6df1186338\") " pod="openshift-marketplace/community-operators-n8d7m" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.249816 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-utilities\") pod \"community-operators-n8d7m\" (UID: \"00f4b4c1-af36-4ccd-a3d1-dd6df1186338\") " pod="openshift-marketplace/community-operators-n8d7m" Feb 20 08:08:31 crc kubenswrapper[4948]: E0220 08:08:31.249893 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:31.749876032 +0000 UTC m=+160.724370852 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.250374 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-catalog-content\") pod \"community-operators-n8d7m\" (UID: \"00f4b4c1-af36-4ccd-a3d1-dd6df1186338\") " pod="openshift-marketplace/community-operators-n8d7m" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.264229 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tvwkh"] Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.268100 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tvwkh" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.294026 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tvwkh"] Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.297872 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxq9k\" (UniqueName: \"kubernetes.io/projected/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-kube-api-access-kxq9k\") pod \"community-operators-n8d7m\" (UID: \"00f4b4c1-af36-4ccd-a3d1-dd6df1186338\") " pod="openshift-marketplace/community-operators-n8d7m" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.315336 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sbf5n" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.350475 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-catalog-content\") pod \"community-operators-tvwkh\" (UID: \"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d\") " pod="openshift-marketplace/community-operators-tvwkh" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.350515 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-utilities\") pod \"community-operators-tvwkh\" (UID: \"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d\") " pod="openshift-marketplace/community-operators-tvwkh" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.350551 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.350592 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px4gp\" (UniqueName: \"kubernetes.io/projected/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-kube-api-access-px4gp\") pod \"community-operators-tvwkh\" (UID: \"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d\") " pod="openshift-marketplace/community-operators-tvwkh" Feb 20 08:08:31 crc kubenswrapper[4948]: E0220 08:08:31.351613 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:31.850963601 +0000 UTC m=+160.825458421 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.423601 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n8d7m" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.452598 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.452981 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-catalog-content\") pod \"community-operators-tvwkh\" (UID: \"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d\") " pod="openshift-marketplace/community-operators-tvwkh" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.453018 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-utilities\") pod \"community-operators-tvwkh\" (UID: \"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d\") " pod="openshift-marketplace/community-operators-tvwkh" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.453072 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px4gp\" (UniqueName: \"kubernetes.io/projected/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-kube-api-access-px4gp\") pod \"community-operators-tvwkh\" (UID: \"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d\") " pod="openshift-marketplace/community-operators-tvwkh" Feb 20 08:08:31 crc kubenswrapper[4948]: E0220 08:08:31.453540 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:31.953522776 +0000 UTC m=+160.928017606 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.454342 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-utilities\") pod \"community-operators-tvwkh\" (UID: \"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d\") " pod="openshift-marketplace/community-operators-tvwkh" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.454388 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-catalog-content\") pod \"community-operators-tvwkh\" (UID: \"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d\") " pod="openshift-marketplace/community-operators-tvwkh" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.473452 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 08:08:31 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Feb 20 08:08:31 crc kubenswrapper[4948]: [+]process-running ok Feb 20 08:08:31 crc kubenswrapper[4948]: healthz check failed Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.473534 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.481013 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nrhl6"] Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.494886 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px4gp\" (UniqueName: \"kubernetes.io/projected/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-kube-api-access-px4gp\") pod \"community-operators-tvwkh\" (UID: \"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d\") " pod="openshift-marketplace/community-operators-tvwkh" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.554901 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:31 crc kubenswrapper[4948]: E0220 08:08:31.557501 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:32.057477515 +0000 UTC m=+161.031972335 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.630335 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tvwkh" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.664728 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:31 crc kubenswrapper[4948]: E0220 08:08:31.665331 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:32.165308919 +0000 UTC m=+161.139803739 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.735399 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sbf5n"] Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.769825 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:31 crc kubenswrapper[4948]: E0220 08:08:31.770224 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:32.270202901 +0000 UTC m=+161.244697721 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.814408 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n8d7m"] Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.853458 4948 generic.go:334] "Generic (PLEG): container finished" podID="f77016a6-cd35-49df-b6b5-65b4858b41c9" containerID="0520ff9e3a8f12a5228a352b59373978c12d5f10bb51923b12eab15acf564086" exitCode=0 Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.854780 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" event={"ID":"f77016a6-cd35-49df-b6b5-65b4858b41c9","Type":"ContainerDied","Data":"0520ff9e3a8f12a5228a352b59373978c12d5f10bb51923b12eab15acf564086"} Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.868121 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-trh7g" event={"ID":"13b4eba0-42e7-448a-998c-d7aaeffc9218","Type":"ContainerStarted","Data":"1f020d0f8752ef48a55c7d3b2aa1b9f3542e54bfa2d906bdd61229fe0bac45d4"} Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.871803 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:31 crc kubenswrapper[4948]: E0220 08:08:31.872152 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:32.372135161 +0000 UTC m=+161.346629981 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.875223 4948 generic.go:334] "Generic (PLEG): container finished" podID="f75a50cf-7dae-420b-a00d-671ede3cb6f7" containerID="df3ff28a9e161bcd3a383de17aefbe231a72f4c519e331260796574dc38d393e" exitCode=0 Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.876785 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nrhl6" event={"ID":"f75a50cf-7dae-420b-a00d-671ede3cb6f7","Type":"ContainerDied","Data":"df3ff28a9e161bcd3a383de17aefbe231a72f4c519e331260796574dc38d393e"} Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.876838 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nrhl6" event={"ID":"f75a50cf-7dae-420b-a00d-671ede3cb6f7","Type":"ContainerStarted","Data":"d0015edf4c699a29515d340ae2c01786195a0e9d51c4459eae81be82178c452a"} Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.894025 4948 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.997201 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:31 crc kubenswrapper[4948]: E0220 08:08:31.998054 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:32.498036159 +0000 UTC m=+161.472530979 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.032312 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-trh7g" podStartSLOduration=11.032288829 podStartE2EDuration="11.032288829s" podCreationTimestamp="2026-02-20 08:08:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:31.999333431 +0000 UTC m=+160.973828251" watchObservedRunningTime="2026-02-20 08:08:32.032288829 +0000 UTC m=+161.006783649" Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.081125 4948 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-02-20T08:08:31.091711343Z","Handler":null,"Name":""} Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.091538 4948 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.091598 4948 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.107738 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.115355 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.209583 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.318381 4948 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.318451 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.359334 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.444186 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tvwkh"] Feb 20 08:08:32 crc kubenswrapper[4948]: W0220 08:08:32.449690 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb6d696cd_f5cf_47e0_af8d_4d20e93bd80d.slice/crio-46adc95f753dc88a78d1ef713a2cbcbf1789c9d4ad7e29e46f1a4a26d4497aeb WatchSource:0}: Error finding container 46adc95f753dc88a78d1ef713a2cbcbf1789c9d4ad7e29e46f1a4a26d4497aeb: Status 404 returned error can't find the container with id 46adc95f753dc88a78d1ef713a2cbcbf1789c9d4ad7e29e46f1a4a26d4497aeb Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.462820 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 08:08:32 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Feb 20 08:08:32 crc kubenswrapper[4948]: [+]process-running ok Feb 20 08:08:32 crc kubenswrapper[4948]: healthz check failed Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.462893 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.594956 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.761658 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.855551 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-44l7j" Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.861715 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-7wm8c"] Feb 20 08:08:32 crc kubenswrapper[4948]: W0220 08:08:32.876422 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57c455f7_aa9c_405c_bee3_89726b84f5db.slice/crio-f90694ba31dc9edaf32c68d5b113799c7c3cf39817bac7a06e98cc5e157c4d16 WatchSource:0}: Error finding container f90694ba31dc9edaf32c68d5b113799c7c3cf39817bac7a06e98cc5e157c4d16: Status 404 returned error can't find the container with id f90694ba31dc9edaf32c68d5b113799c7c3cf39817bac7a06e98cc5e157c4d16 Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.884240 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tvwkh" event={"ID":"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d","Type":"ContainerStarted","Data":"46adc95f753dc88a78d1ef713a2cbcbf1789c9d4ad7e29e46f1a4a26d4497aeb"} Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.886338 4948 generic.go:334] "Generic (PLEG): container finished" podID="000032f5-8df9-4879-8a8c-7ca06deb0862" containerID="b939d98be77246f833198ae82a9cce74eaf753fb0b14e37eb3663fc9e452cd93" exitCode=0 Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.886787 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sbf5n" event={"ID":"000032f5-8df9-4879-8a8c-7ca06deb0862","Type":"ContainerDied","Data":"b939d98be77246f833198ae82a9cce74eaf753fb0b14e37eb3663fc9e452cd93"} Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.886831 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sbf5n" event={"ID":"000032f5-8df9-4879-8a8c-7ca06deb0862","Type":"ContainerStarted","Data":"6682a9f05af301b707c145d9ce47468c3451f6e6fc6b07c7b78c99356b765594"} Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.897249 4948 generic.go:334] "Generic (PLEG): container finished" podID="00f4b4c1-af36-4ccd-a3d1-dd6df1186338" containerID="8ae4d6ef7b64ba80a0da276ad7998550243e3a26aaa0bbf3282f22cf4366103f" exitCode=0 Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.898243 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n8d7m" event={"ID":"00f4b4c1-af36-4ccd-a3d1-dd6df1186338","Type":"ContainerDied","Data":"8ae4d6ef7b64ba80a0da276ad7998550243e3a26aaa0bbf3282f22cf4366103f"} Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.898359 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n8d7m" event={"ID":"00f4b4c1-af36-4ccd-a3d1-dd6df1186338","Type":"ContainerStarted","Data":"33b053c8152742a922c7d8709831de18b8993625ae7594197966e98f173b1174"} Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.899246 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.900145 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.908602 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.909588 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.909770 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.024936 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b189bd1e-6355-433c-972d-2e27f98fe153-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b189bd1e-6355-433c-972d-2e27f98fe153\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.025014 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b189bd1e-6355-433c-972d-2e27f98fe153-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b189bd1e-6355-433c-972d-2e27f98fe153\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.090181 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c9z42"] Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.095167 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c9z42" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.097898 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.129299 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b189bd1e-6355-433c-972d-2e27f98fe153-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b189bd1e-6355-433c-972d-2e27f98fe153\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.129236 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b189bd1e-6355-433c-972d-2e27f98fe153-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b189bd1e-6355-433c-972d-2e27f98fe153\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.130672 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b189bd1e-6355-433c-972d-2e27f98fe153-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b189bd1e-6355-433c-972d-2e27f98fe153\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.149936 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c9z42"] Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.171733 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b189bd1e-6355-433c-972d-2e27f98fe153-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b189bd1e-6355-433c-972d-2e27f98fe153\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.173457 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.181462 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.181488 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.201182 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.233652 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-utilities\") pod \"redhat-marketplace-c9z42\" (UID: \"c53ddd33-0f9f-4794-b346-7d48a6c09c9b\") " pod="openshift-marketplace/redhat-marketplace-c9z42" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.233714 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqtvv\" (UniqueName: \"kubernetes.io/projected/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-kube-api-access-cqtvv\") pod \"redhat-marketplace-c9z42\" (UID: \"c53ddd33-0f9f-4794-b346-7d48a6c09c9b\") " pod="openshift-marketplace/redhat-marketplace-c9z42" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.233770 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-catalog-content\") pod \"redhat-marketplace-c9z42\" (UID: \"c53ddd33-0f9f-4794-b346-7d48a6c09c9b\") " pod="openshift-marketplace/redhat-marketplace-c9z42" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.324656 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.335272 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f77016a6-cd35-49df-b6b5-65b4858b41c9-secret-volume\") pod \"f77016a6-cd35-49df-b6b5-65b4858b41c9\" (UID: \"f77016a6-cd35-49df-b6b5-65b4858b41c9\") " Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.335360 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ql4r\" (UniqueName: \"kubernetes.io/projected/f77016a6-cd35-49df-b6b5-65b4858b41c9-kube-api-access-2ql4r\") pod \"f77016a6-cd35-49df-b6b5-65b4858b41c9\" (UID: \"f77016a6-cd35-49df-b6b5-65b4858b41c9\") " Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.335417 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f77016a6-cd35-49df-b6b5-65b4858b41c9-config-volume\") pod \"f77016a6-cd35-49df-b6b5-65b4858b41c9\" (UID: \"f77016a6-cd35-49df-b6b5-65b4858b41c9\") " Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.335686 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqtvv\" (UniqueName: \"kubernetes.io/projected/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-kube-api-access-cqtvv\") pod \"redhat-marketplace-c9z42\" (UID: \"c53ddd33-0f9f-4794-b346-7d48a6c09c9b\") " pod="openshift-marketplace/redhat-marketplace-c9z42" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.335805 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-catalog-content\") pod \"redhat-marketplace-c9z42\" (UID: \"c53ddd33-0f9f-4794-b346-7d48a6c09c9b\") " pod="openshift-marketplace/redhat-marketplace-c9z42" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.335958 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-utilities\") pod \"redhat-marketplace-c9z42\" (UID: \"c53ddd33-0f9f-4794-b346-7d48a6c09c9b\") " pod="openshift-marketplace/redhat-marketplace-c9z42" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.336698 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-utilities\") pod \"redhat-marketplace-c9z42\" (UID: \"c53ddd33-0f9f-4794-b346-7d48a6c09c9b\") " pod="openshift-marketplace/redhat-marketplace-c9z42" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.338718 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-catalog-content\") pod \"redhat-marketplace-c9z42\" (UID: \"c53ddd33-0f9f-4794-b346-7d48a6c09c9b\") " pod="openshift-marketplace/redhat-marketplace-c9z42" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.341254 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f77016a6-cd35-49df-b6b5-65b4858b41c9-config-volume" (OuterVolumeSpecName: "config-volume") pod "f77016a6-cd35-49df-b6b5-65b4858b41c9" (UID: "f77016a6-cd35-49df-b6b5-65b4858b41c9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.341746 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f77016a6-cd35-49df-b6b5-65b4858b41c9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f77016a6-cd35-49df-b6b5-65b4858b41c9" (UID: "f77016a6-cd35-49df-b6b5-65b4858b41c9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.342745 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f77016a6-cd35-49df-b6b5-65b4858b41c9-kube-api-access-2ql4r" (OuterVolumeSpecName: "kube-api-access-2ql4r") pod "f77016a6-cd35-49df-b6b5-65b4858b41c9" (UID: "f77016a6-cd35-49df-b6b5-65b4858b41c9"). InnerVolumeSpecName "kube-api-access-2ql4r". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.357224 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqtvv\" (UniqueName: \"kubernetes.io/projected/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-kube-api-access-cqtvv\") pod \"redhat-marketplace-c9z42\" (UID: \"c53ddd33-0f9f-4794-b346-7d48a6c09c9b\") " pod="openshift-marketplace/redhat-marketplace-c9z42" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.422109 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c9z42" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.436930 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ql4r\" (UniqueName: \"kubernetes.io/projected/f77016a6-cd35-49df-b6b5-65b4858b41c9-kube-api-access-2ql4r\") on node \"crc\" DevicePath \"\"" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.436955 4948 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f77016a6-cd35-49df-b6b5-65b4858b41c9-config-volume\") on node \"crc\" DevicePath \"\"" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.436978 4948 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f77016a6-cd35-49df-b6b5-65b4858b41c9-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.467883 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 08:08:33 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Feb 20 08:08:33 crc kubenswrapper[4948]: [+]process-running ok Feb 20 08:08:33 crc kubenswrapper[4948]: healthz check failed Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.467948 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.469130 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-67ptx"] Feb 20 08:08:33 crc kubenswrapper[4948]: E0220 08:08:33.469919 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f77016a6-cd35-49df-b6b5-65b4858b41c9" containerName="collect-profiles" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.469939 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f77016a6-cd35-49df-b6b5-65b4858b41c9" containerName="collect-profiles" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.470042 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f77016a6-cd35-49df-b6b5-65b4858b41c9" containerName="collect-profiles" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.472247 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-67ptx" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.479262 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-67ptx"] Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.579247 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 20 08:08:33 crc kubenswrapper[4948]: W0220 08:08:33.592343 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podb189bd1e_6355_433c_972d_2e27f98fe153.slice/crio-df6078b5b4a4366ce4b18c7e57e97fa910c6f7c0860321f7de0754fed2628438 WatchSource:0}: Error finding container df6078b5b4a4366ce4b18c7e57e97fa910c6f7c0860321f7de0754fed2628438: Status 404 returned error can't find the container with id df6078b5b4a4366ce4b18c7e57e97fa910c6f7c0860321f7de0754fed2628438 Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.639874 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8150552-4914-4e92-b041-56600ed6d2f5-utilities\") pod \"redhat-marketplace-67ptx\" (UID: \"b8150552-4914-4e92-b041-56600ed6d2f5\") " pod="openshift-marketplace/redhat-marketplace-67ptx" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.639918 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8150552-4914-4e92-b041-56600ed6d2f5-catalog-content\") pod \"redhat-marketplace-67ptx\" (UID: \"b8150552-4914-4e92-b041-56600ed6d2f5\") " pod="openshift-marketplace/redhat-marketplace-67ptx" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.640136 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cx26r\" (UniqueName: \"kubernetes.io/projected/b8150552-4914-4e92-b041-56600ed6d2f5-kube-api-access-cx26r\") pod \"redhat-marketplace-67ptx\" (UID: \"b8150552-4914-4e92-b041-56600ed6d2f5\") " pod="openshift-marketplace/redhat-marketplace-67ptx" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.664909 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c9z42"] Feb 20 08:08:33 crc kubenswrapper[4948]: W0220 08:08:33.677193 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc53ddd33_0f9f_4794_b346_7d48a6c09c9b.slice/crio-9a7c5f1538f439aba0332cfbd20f5f653d7b240472d961eff61974598f86365b WatchSource:0}: Error finding container 9a7c5f1538f439aba0332cfbd20f5f653d7b240472d961eff61974598f86365b: Status 404 returned error can't find the container with id 9a7c5f1538f439aba0332cfbd20f5f653d7b240472d961eff61974598f86365b Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.733557 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.741466 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8150552-4914-4e92-b041-56600ed6d2f5-utilities\") pod \"redhat-marketplace-67ptx\" (UID: \"b8150552-4914-4e92-b041-56600ed6d2f5\") " pod="openshift-marketplace/redhat-marketplace-67ptx" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.741516 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8150552-4914-4e92-b041-56600ed6d2f5-catalog-content\") pod \"redhat-marketplace-67ptx\" (UID: \"b8150552-4914-4e92-b041-56600ed6d2f5\") " pod="openshift-marketplace/redhat-marketplace-67ptx" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.741630 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cx26r\" (UniqueName: \"kubernetes.io/projected/b8150552-4914-4e92-b041-56600ed6d2f5-kube-api-access-cx26r\") pod \"redhat-marketplace-67ptx\" (UID: \"b8150552-4914-4e92-b041-56600ed6d2f5\") " pod="openshift-marketplace/redhat-marketplace-67ptx" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.742567 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8150552-4914-4e92-b041-56600ed6d2f5-utilities\") pod \"redhat-marketplace-67ptx\" (UID: \"b8150552-4914-4e92-b041-56600ed6d2f5\") " pod="openshift-marketplace/redhat-marketplace-67ptx" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.742612 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8150552-4914-4e92-b041-56600ed6d2f5-catalog-content\") pod \"redhat-marketplace-67ptx\" (UID: \"b8150552-4914-4e92-b041-56600ed6d2f5\") " pod="openshift-marketplace/redhat-marketplace-67ptx" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.765782 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cx26r\" (UniqueName: \"kubernetes.io/projected/b8150552-4914-4e92-b041-56600ed6d2f5-kube-api-access-cx26r\") pod \"redhat-marketplace-67ptx\" (UID: \"b8150552-4914-4e92-b041-56600ed6d2f5\") " pod="openshift-marketplace/redhat-marketplace-67ptx" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.777842 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.777903 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.785853 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.794484 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-67ptx" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.818243 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.818311 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.820253 4948 patch_prober.go:28] interesting pod/console-f9d7485db-k8zcr container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.820360 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-k8zcr" podUID="f81366e5-fbdb-4a0b-a91f-f603d9b8570d" containerName="console" probeResult="failure" output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.858606 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-frznz"] Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.859893 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-frznz" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.861894 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.878192 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-frznz"] Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.939935 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9z42" event={"ID":"c53ddd33-0f9f-4794-b346-7d48a6c09c9b","Type":"ContainerStarted","Data":"1d07ae65c7f4887c7cc6c699e956a80f175ba6deadc7da37863658727e044dbe"} Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.940063 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9z42" event={"ID":"c53ddd33-0f9f-4794-b346-7d48a6c09c9b","Type":"ContainerStarted","Data":"9a7c5f1538f439aba0332cfbd20f5f653d7b240472d961eff61974598f86365b"} Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.944480 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.944474 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" event={"ID":"f77016a6-cd35-49df-b6b5-65b4858b41c9","Type":"ContainerDied","Data":"92efa091cdddd802dc78aa2d9bf128d9d0f0ac869c640718f1c1a8d1aead053f"} Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.944892 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92efa091cdddd802dc78aa2d9bf128d9d0f0ac869c640718f1c1a8d1aead053f" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.945440 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-utilities\") pod \"redhat-operators-frznz\" (UID: \"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9\") " pod="openshift-marketplace/redhat-operators-frznz" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.945626 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-catalog-content\") pod \"redhat-operators-frznz\" (UID: \"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9\") " pod="openshift-marketplace/redhat-operators-frznz" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.945789 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqs4x\" (UniqueName: \"kubernetes.io/projected/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-kube-api-access-hqs4x\") pod \"redhat-operators-frznz\" (UID: \"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9\") " pod="openshift-marketplace/redhat-operators-frznz" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.947686 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" event={"ID":"57c455f7-aa9c-405c-bee3-89726b84f5db","Type":"ContainerStarted","Data":"20908318b48f1c9de151f5bb5f43a14fb60c5d750ea49471a9d574dabed15120"} Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.947742 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" event={"ID":"57c455f7-aa9c-405c-bee3-89726b84f5db","Type":"ContainerStarted","Data":"f90694ba31dc9edaf32c68d5b113799c7c3cf39817bac7a06e98cc5e157c4d16"} Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.948106 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.950361 4948 generic.go:334] "Generic (PLEG): container finished" podID="b6d696cd-f5cf-47e0-af8d-4d20e93bd80d" containerID="5e2b81703211ae6dafce6a9ca7e54224c7be8c36c969c7e7bf86ed1763e784b1" exitCode=0 Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.950442 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tvwkh" event={"ID":"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d","Type":"ContainerDied","Data":"5e2b81703211ae6dafce6a9ca7e54224c7be8c36c969c7e7bf86ed1763e784b1"} Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.958607 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b189bd1e-6355-433c-972d-2e27f98fe153","Type":"ContainerStarted","Data":"df6078b5b4a4366ce4b18c7e57e97fa910c6f7c0860321f7de0754fed2628438"} Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.966267 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.968846 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.991654 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" podStartSLOduration=139.991593848 podStartE2EDuration="2m19.991593848s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:33.987680922 +0000 UTC m=+162.962175742" watchObservedRunningTime="2026-02-20 08:08:33.991593848 +0000 UTC m=+162.966088668" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.049639 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqs4x\" (UniqueName: \"kubernetes.io/projected/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-kube-api-access-hqs4x\") pod \"redhat-operators-frznz\" (UID: \"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9\") " pod="openshift-marketplace/redhat-operators-frznz" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.049734 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-utilities\") pod \"redhat-operators-frznz\" (UID: \"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9\") " pod="openshift-marketplace/redhat-operators-frznz" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.049855 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-catalog-content\") pod \"redhat-operators-frznz\" (UID: \"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9\") " pod="openshift-marketplace/redhat-operators-frznz" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.055556 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-catalog-content\") pod \"redhat-operators-frznz\" (UID: \"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9\") " pod="openshift-marketplace/redhat-operators-frznz" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.058045 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-utilities\") pod \"redhat-operators-frznz\" (UID: \"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9\") " pod="openshift-marketplace/redhat-operators-frznz" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.111451 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-snc6x"] Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.112780 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-snc6x" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.114904 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqs4x\" (UniqueName: \"kubernetes.io/projected/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-kube-api-access-hqs4x\") pod \"redhat-operators-frznz\" (UID: \"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9\") " pod="openshift-marketplace/redhat-operators-frznz" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.125033 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-snc6x"] Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.169031 4948 patch_prober.go:28] interesting pod/downloads-7954f5f757-grnfc container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.169400 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-grnfc" podUID="bc13d9fe-fd58-4d3b-9278-4f3e59da5976" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.180203 4948 patch_prober.go:28] interesting pod/downloads-7954f5f757-grnfc container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.180279 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-grnfc" podUID="bc13d9fe-fd58-4d3b-9278-4f3e59da5976" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.215395 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-frznz" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.253109 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5ec9802-aa9c-4518-b99a-1673ab75ec50-catalog-content\") pod \"redhat-operators-snc6x\" (UID: \"b5ec9802-aa9c-4518-b99a-1673ab75ec50\") " pod="openshift-marketplace/redhat-operators-snc6x" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.253181 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gc4n5\" (UniqueName: \"kubernetes.io/projected/b5ec9802-aa9c-4518-b99a-1673ab75ec50-kube-api-access-gc4n5\") pod \"redhat-operators-snc6x\" (UID: \"b5ec9802-aa9c-4518-b99a-1673ab75ec50\") " pod="openshift-marketplace/redhat-operators-snc6x" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.253266 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5ec9802-aa9c-4518-b99a-1673ab75ec50-utilities\") pod \"redhat-operators-snc6x\" (UID: \"b5ec9802-aa9c-4518-b99a-1673ab75ec50\") " pod="openshift-marketplace/redhat-operators-snc6x" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.351365 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-67ptx"] Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.361965 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5ec9802-aa9c-4518-b99a-1673ab75ec50-utilities\") pod \"redhat-operators-snc6x\" (UID: \"b5ec9802-aa9c-4518-b99a-1673ab75ec50\") " pod="openshift-marketplace/redhat-operators-snc6x" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.362094 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5ec9802-aa9c-4518-b99a-1673ab75ec50-catalog-content\") pod \"redhat-operators-snc6x\" (UID: \"b5ec9802-aa9c-4518-b99a-1673ab75ec50\") " pod="openshift-marketplace/redhat-operators-snc6x" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.362134 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gc4n5\" (UniqueName: \"kubernetes.io/projected/b5ec9802-aa9c-4518-b99a-1673ab75ec50-kube-api-access-gc4n5\") pod \"redhat-operators-snc6x\" (UID: \"b5ec9802-aa9c-4518-b99a-1673ab75ec50\") " pod="openshift-marketplace/redhat-operators-snc6x" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.362449 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5ec9802-aa9c-4518-b99a-1673ab75ec50-utilities\") pod \"redhat-operators-snc6x\" (UID: \"b5ec9802-aa9c-4518-b99a-1673ab75ec50\") " pod="openshift-marketplace/redhat-operators-snc6x" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.362729 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5ec9802-aa9c-4518-b99a-1673ab75ec50-catalog-content\") pod \"redhat-operators-snc6x\" (UID: \"b5ec9802-aa9c-4518-b99a-1673ab75ec50\") " pod="openshift-marketplace/redhat-operators-snc6x" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.430775 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gc4n5\" (UniqueName: \"kubernetes.io/projected/b5ec9802-aa9c-4518-b99a-1673ab75ec50-kube-api-access-gc4n5\") pod \"redhat-operators-snc6x\" (UID: \"b5ec9802-aa9c-4518-b99a-1673ab75ec50\") " pod="openshift-marketplace/redhat-operators-snc6x" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.465632 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.466067 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-snc6x" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.471438 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 08:08:34 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Feb 20 08:08:34 crc kubenswrapper[4948]: [+]process-running ok Feb 20 08:08:34 crc kubenswrapper[4948]: healthz check failed Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.471489 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.625395 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-frznz"] Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.630915 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.875555 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-snc6x"] Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.011070 4948 generic.go:334] "Generic (PLEG): container finished" podID="dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" containerID="6e6f9609316227fbb42f161491bc563af5bbab43a8ee401a26665bf588b0e5d7" exitCode=0 Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.011275 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frznz" event={"ID":"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9","Type":"ContainerDied","Data":"6e6f9609316227fbb42f161491bc563af5bbab43a8ee401a26665bf588b0e5d7"} Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.011537 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frznz" event={"ID":"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9","Type":"ContainerStarted","Data":"bc370df4b0b552966157b6ac4af296601fb860ed2508da9c4291e9924ec1e9cb"} Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.022917 4948 generic.go:334] "Generic (PLEG): container finished" podID="c53ddd33-0f9f-4794-b346-7d48a6c09c9b" containerID="1d07ae65c7f4887c7cc6c699e956a80f175ba6deadc7da37863658727e044dbe" exitCode=0 Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.023000 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9z42" event={"ID":"c53ddd33-0f9f-4794-b346-7d48a6c09c9b","Type":"ContainerDied","Data":"1d07ae65c7f4887c7cc6c699e956a80f175ba6deadc7da37863658727e044dbe"} Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.036600 4948 generic.go:334] "Generic (PLEG): container finished" podID="b8150552-4914-4e92-b041-56600ed6d2f5" containerID="24232a1b4e66920a8c674176ee38da24ffd8d17de7ae8a51e030daddaaeb552d" exitCode=0 Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.036709 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-67ptx" event={"ID":"b8150552-4914-4e92-b041-56600ed6d2f5","Type":"ContainerDied","Data":"24232a1b4e66920a8c674176ee38da24ffd8d17de7ae8a51e030daddaaeb552d"} Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.037549 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-67ptx" event={"ID":"b8150552-4914-4e92-b041-56600ed6d2f5","Type":"ContainerStarted","Data":"b740ba77241ab7ac59567dc96b03935ff231d4a2d8855de4a6ed59ddebaef76b"} Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.048551 4948 generic.go:334] "Generic (PLEG): container finished" podID="b189bd1e-6355-433c-972d-2e27f98fe153" containerID="e48b6bdd3e78b89a1786debf78466f80dde9962619930bae2b9cae797f181fd2" exitCode=0 Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.050675 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b189bd1e-6355-433c-972d-2e27f98fe153","Type":"ContainerDied","Data":"e48b6bdd3e78b89a1786debf78466f80dde9962619930bae2b9cae797f181fd2"} Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.219083 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.219939 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.224748 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.224748 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.226090 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.293136 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.293269 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.395367 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.395457 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.395855 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.420534 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.470340 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 08:08:35 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Feb 20 08:08:35 crc kubenswrapper[4948]: [+]process-running ok Feb 20 08:08:35 crc kubenswrapper[4948]: healthz check failed Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.470409 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.552308 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.916905 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 20 08:08:35 crc kubenswrapper[4948]: W0220 08:08:35.969608 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podc2d35955_eb5e_4d11_8e5a_5ed0716a5bc0.slice/crio-b80e153205aa2bfb003f43118045cec88da3439901cf54d41fc4f1ecf766667b WatchSource:0}: Error finding container b80e153205aa2bfb003f43118045cec88da3439901cf54d41fc4f1ecf766667b: Status 404 returned error can't find the container with id b80e153205aa2bfb003f43118045cec88da3439901cf54d41fc4f1ecf766667b Feb 20 08:08:36 crc kubenswrapper[4948]: I0220 08:08:36.056921 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0","Type":"ContainerStarted","Data":"b80e153205aa2bfb003f43118045cec88da3439901cf54d41fc4f1ecf766667b"} Feb 20 08:08:36 crc kubenswrapper[4948]: I0220 08:08:36.059280 4948 generic.go:334] "Generic (PLEG): container finished" podID="b5ec9802-aa9c-4518-b99a-1673ab75ec50" containerID="5e7272299ae3bd4fd920b69a3700514cf14947047239ae7f3143eafd6f2d9320" exitCode=0 Feb 20 08:08:36 crc kubenswrapper[4948]: I0220 08:08:36.059433 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-snc6x" event={"ID":"b5ec9802-aa9c-4518-b99a-1673ab75ec50","Type":"ContainerDied","Data":"5e7272299ae3bd4fd920b69a3700514cf14947047239ae7f3143eafd6f2d9320"} Feb 20 08:08:36 crc kubenswrapper[4948]: I0220 08:08:36.059527 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-snc6x" event={"ID":"b5ec9802-aa9c-4518-b99a-1673ab75ec50","Type":"ContainerStarted","Data":"40f046d646c178228156ce2281e80f2e2be9092681446e798f497080898603ae"} Feb 20 08:08:36 crc kubenswrapper[4948]: I0220 08:08:36.457846 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 08:08:36 crc kubenswrapper[4948]: I0220 08:08:36.464839 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 08:08:36 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Feb 20 08:08:36 crc kubenswrapper[4948]: [+]process-running ok Feb 20 08:08:36 crc kubenswrapper[4948]: healthz check failed Feb 20 08:08:36 crc kubenswrapper[4948]: I0220 08:08:36.464904 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:08:36 crc kubenswrapper[4948]: I0220 08:08:36.522413 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b189bd1e-6355-433c-972d-2e27f98fe153-kube-api-access\") pod \"b189bd1e-6355-433c-972d-2e27f98fe153\" (UID: \"b189bd1e-6355-433c-972d-2e27f98fe153\") " Feb 20 08:08:36 crc kubenswrapper[4948]: I0220 08:08:36.522510 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b189bd1e-6355-433c-972d-2e27f98fe153-kubelet-dir\") pod \"b189bd1e-6355-433c-972d-2e27f98fe153\" (UID: \"b189bd1e-6355-433c-972d-2e27f98fe153\") " Feb 20 08:08:36 crc kubenswrapper[4948]: I0220 08:08:36.522792 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b189bd1e-6355-433c-972d-2e27f98fe153-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "b189bd1e-6355-433c-972d-2e27f98fe153" (UID: "b189bd1e-6355-433c-972d-2e27f98fe153"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:08:36 crc kubenswrapper[4948]: I0220 08:08:36.541478 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b189bd1e-6355-433c-972d-2e27f98fe153-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "b189bd1e-6355-433c-972d-2e27f98fe153" (UID: "b189bd1e-6355-433c-972d-2e27f98fe153"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:08:36 crc kubenswrapper[4948]: I0220 08:08:36.624383 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b189bd1e-6355-433c-972d-2e27f98fe153-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 20 08:08:36 crc kubenswrapper[4948]: I0220 08:08:36.624439 4948 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b189bd1e-6355-433c-972d-2e27f98fe153-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 20 08:08:36 crc kubenswrapper[4948]: I0220 08:08:36.725636 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs\") pod \"network-metrics-daemon-xd86g\" (UID: \"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\") " pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:08:36 crc kubenswrapper[4948]: I0220 08:08:36.752756 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs\") pod \"network-metrics-daemon-xd86g\" (UID: \"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\") " pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:08:36 crc kubenswrapper[4948]: I0220 08:08:36.868170 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:08:37 crc kubenswrapper[4948]: I0220 08:08:37.135666 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0","Type":"ContainerStarted","Data":"e57316aee001d124820060d888331a36dba4f8f0e8e53102f7df766f39d91d8a"} Feb 20 08:08:37 crc kubenswrapper[4948]: I0220 08:08:37.149781 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b189bd1e-6355-433c-972d-2e27f98fe153","Type":"ContainerDied","Data":"df6078b5b4a4366ce4b18c7e57e97fa910c6f7c0860321f7de0754fed2628438"} Feb 20 08:08:37 crc kubenswrapper[4948]: I0220 08:08:37.149836 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df6078b5b4a4366ce4b18c7e57e97fa910c6f7c0860321f7de0754fed2628438" Feb 20 08:08:37 crc kubenswrapper[4948]: I0220 08:08:37.149875 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 08:08:37 crc kubenswrapper[4948]: I0220 08:08:37.158610 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.158588624 podStartE2EDuration="2.158588624s" podCreationTimestamp="2026-02-20 08:08:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:37.15391089 +0000 UTC m=+166.128405710" watchObservedRunningTime="2026-02-20 08:08:37.158588624 +0000 UTC m=+166.133083444" Feb 20 08:08:37 crc kubenswrapper[4948]: I0220 08:08:37.252680 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-xd86g"] Feb 20 08:08:37 crc kubenswrapper[4948]: I0220 08:08:37.466400 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 08:08:37 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Feb 20 08:08:37 crc kubenswrapper[4948]: [+]process-running ok Feb 20 08:08:37 crc kubenswrapper[4948]: healthz check failed Feb 20 08:08:37 crc kubenswrapper[4948]: I0220 08:08:37.466502 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:08:38 crc kubenswrapper[4948]: I0220 08:08:38.025370 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:08:38 crc kubenswrapper[4948]: I0220 08:08:38.025687 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:08:38 crc kubenswrapper[4948]: I0220 08:08:38.174110 4948 generic.go:334] "Generic (PLEG): container finished" podID="c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0" containerID="e57316aee001d124820060d888331a36dba4f8f0e8e53102f7df766f39d91d8a" exitCode=0 Feb 20 08:08:38 crc kubenswrapper[4948]: I0220 08:08:38.174174 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0","Type":"ContainerDied","Data":"e57316aee001d124820060d888331a36dba4f8f0e8e53102f7df766f39d91d8a"} Feb 20 08:08:38 crc kubenswrapper[4948]: I0220 08:08:38.203381 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-xd86g" event={"ID":"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd","Type":"ContainerStarted","Data":"b1f89ee38d3a256c46d4126a49d66b2ee3a928d4035b624bbc59430d1a8c8dff"} Feb 20 08:08:38 crc kubenswrapper[4948]: I0220 08:08:38.461543 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 08:08:38 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Feb 20 08:08:38 crc kubenswrapper[4948]: [+]process-running ok Feb 20 08:08:38 crc kubenswrapper[4948]: healthz check failed Feb 20 08:08:38 crc kubenswrapper[4948]: I0220 08:08:38.461620 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:08:39 crc kubenswrapper[4948]: I0220 08:08:39.224524 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-xd86g" event={"ID":"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd","Type":"ContainerStarted","Data":"027950bf79b0645503bcffbc775f0c1cffa721ad8382c116eab7d302c5f8f2a2"} Feb 20 08:08:39 crc kubenswrapper[4948]: I0220 08:08:39.225017 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-xd86g" event={"ID":"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd","Type":"ContainerStarted","Data":"6bbc067891d2ea7f9b69d2c5053c14eecb4759d3497f466543ffce6b10dd81e9"} Feb 20 08:08:39 crc kubenswrapper[4948]: I0220 08:08:39.263680 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-xd86g" podStartSLOduration=145.263651007 podStartE2EDuration="2m25.263651007s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:39.242065548 +0000 UTC m=+168.216560368" watchObservedRunningTime="2026-02-20 08:08:39.263651007 +0000 UTC m=+168.238145827" Feb 20 08:08:39 crc kubenswrapper[4948]: I0220 08:08:39.466701 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 08:08:39 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Feb 20 08:08:39 crc kubenswrapper[4948]: [+]process-running ok Feb 20 08:08:39 crc kubenswrapper[4948]: healthz check failed Feb 20 08:08:39 crc kubenswrapper[4948]: I0220 08:08:39.466766 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:08:39 crc kubenswrapper[4948]: I0220 08:08:39.633729 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 08:08:39 crc kubenswrapper[4948]: I0220 08:08:39.652886 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-9ls2l" Feb 20 08:08:39 crc kubenswrapper[4948]: I0220 08:08:39.713317 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0-kube-api-access\") pod \"c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0\" (UID: \"c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0\") " Feb 20 08:08:39 crc kubenswrapper[4948]: I0220 08:08:39.713492 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0-kubelet-dir\") pod \"c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0\" (UID: \"c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0\") " Feb 20 08:08:39 crc kubenswrapper[4948]: I0220 08:08:39.713920 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0" (UID: "c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:08:39 crc kubenswrapper[4948]: I0220 08:08:39.715368 4948 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 20 08:08:39 crc kubenswrapper[4948]: I0220 08:08:39.739061 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0" (UID: "c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:08:39 crc kubenswrapper[4948]: I0220 08:08:39.816335 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 20 08:08:40 crc kubenswrapper[4948]: I0220 08:08:40.269611 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 08:08:40 crc kubenswrapper[4948]: I0220 08:08:40.270006 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0","Type":"ContainerDied","Data":"b80e153205aa2bfb003f43118045cec88da3439901cf54d41fc4f1ecf766667b"} Feb 20 08:08:40 crc kubenswrapper[4948]: I0220 08:08:40.270042 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b80e153205aa2bfb003f43118045cec88da3439901cf54d41fc4f1ecf766667b" Feb 20 08:08:40 crc kubenswrapper[4948]: I0220 08:08:40.462516 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 08:08:40 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Feb 20 08:08:40 crc kubenswrapper[4948]: [+]process-running ok Feb 20 08:08:40 crc kubenswrapper[4948]: healthz check failed Feb 20 08:08:40 crc kubenswrapper[4948]: I0220 08:08:40.462600 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:08:41 crc kubenswrapper[4948]: I0220 08:08:41.463876 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 08:08:41 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Feb 20 08:08:41 crc kubenswrapper[4948]: [+]process-running ok Feb 20 08:08:41 crc kubenswrapper[4948]: healthz check failed Feb 20 08:08:41 crc kubenswrapper[4948]: I0220 08:08:41.463953 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:08:42 crc kubenswrapper[4948]: I0220 08:08:42.462190 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 08:08:42 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Feb 20 08:08:42 crc kubenswrapper[4948]: [+]process-running ok Feb 20 08:08:42 crc kubenswrapper[4948]: healthz check failed Feb 20 08:08:42 crc kubenswrapper[4948]: I0220 08:08:42.462272 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:08:43 crc kubenswrapper[4948]: I0220 08:08:43.462214 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 08:08:43 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Feb 20 08:08:43 crc kubenswrapper[4948]: [+]process-running ok Feb 20 08:08:43 crc kubenswrapper[4948]: healthz check failed Feb 20 08:08:43 crc kubenswrapper[4948]: I0220 08:08:43.462567 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:08:43 crc kubenswrapper[4948]: I0220 08:08:43.817123 4948 patch_prober.go:28] interesting pod/console-f9d7485db-k8zcr container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Feb 20 08:08:43 crc kubenswrapper[4948]: I0220 08:08:43.817204 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-k8zcr" podUID="f81366e5-fbdb-4a0b-a91f-f603d9b8570d" containerName="console" probeResult="failure" output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" Feb 20 08:08:44 crc kubenswrapper[4948]: I0220 08:08:44.176380 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-grnfc" Feb 20 08:08:44 crc kubenswrapper[4948]: I0220 08:08:44.462103 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 08:08:44 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Feb 20 08:08:44 crc kubenswrapper[4948]: [+]process-running ok Feb 20 08:08:44 crc kubenswrapper[4948]: healthz check failed Feb 20 08:08:44 crc kubenswrapper[4948]: I0220 08:08:44.462187 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:08:45 crc kubenswrapper[4948]: I0220 08:08:45.463167 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:45 crc kubenswrapper[4948]: I0220 08:08:45.470385 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:52 crc kubenswrapper[4948]: I0220 08:08:52.599644 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:53 crc kubenswrapper[4948]: I0220 08:08:53.822519 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:53 crc kubenswrapper[4948]: I0220 08:08:53.826827 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:57 crc kubenswrapper[4948]: E0220 08:08:57.182520 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 20 08:08:57 crc kubenswrapper[4948]: E0220 08:08:57.184354 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nc669,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-nrhl6_openshift-marketplace(f75a50cf-7dae-420b-a00d-671ede3cb6f7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 20 08:08:57 crc kubenswrapper[4948]: E0220 08:08:57.185664 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-nrhl6" podUID="f75a50cf-7dae-420b-a00d-671ede3cb6f7" Feb 20 08:08:58 crc kubenswrapper[4948]: E0220 08:08:58.716240 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-nrhl6" podUID="f75a50cf-7dae-420b-a00d-671ede3cb6f7" Feb 20 08:09:00 crc kubenswrapper[4948]: I0220 08:09:00.965278 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:09:04 crc kubenswrapper[4948]: I0220 08:09:04.275908 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-46sg8" Feb 20 08:09:05 crc kubenswrapper[4948]: E0220 08:09:05.627690 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 20 08:09:05 crc kubenswrapper[4948]: E0220 08:09:05.628342 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gc4n5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-snc6x_openshift-marketplace(b5ec9802-aa9c-4518-b99a-1673ab75ec50): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 20 08:09:05 crc kubenswrapper[4948]: E0220 08:09:05.629575 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-snc6x" podUID="b5ec9802-aa9c-4518-b99a-1673ab75ec50" Feb 20 08:09:05 crc kubenswrapper[4948]: E0220 08:09:05.690490 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 20 08:09:05 crc kubenswrapper[4948]: E0220 08:09:05.690719 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hqs4x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-frznz_openshift-marketplace(dc2fe1f3-084a-4e45-b64d-f385fb28d7e9): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 20 08:09:05 crc kubenswrapper[4948]: E0220 08:09:05.691887 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-frznz" podUID="dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" Feb 20 08:09:05 crc kubenswrapper[4948]: E0220 08:09:05.711226 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 20 08:09:05 crc kubenswrapper[4948]: E0220 08:09:05.711396 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cx26r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-67ptx_openshift-marketplace(b8150552-4914-4e92-b041-56600ed6d2f5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 20 08:09:05 crc kubenswrapper[4948]: E0220 08:09:05.712696 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-67ptx" podUID="b8150552-4914-4e92-b041-56600ed6d2f5" Feb 20 08:09:05 crc kubenswrapper[4948]: E0220 08:09:05.713367 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 20 08:09:05 crc kubenswrapper[4948]: E0220 08:09:05.713529 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-px4gp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-tvwkh_openshift-marketplace(b6d696cd-f5cf-47e0-af8d-4d20e93bd80d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 20 08:09:05 crc kubenswrapper[4948]: E0220 08:09:05.714696 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-tvwkh" podUID="b6d696cd-f5cf-47e0-af8d-4d20e93bd80d" Feb 20 08:09:06 crc kubenswrapper[4948]: I0220 08:09:06.477987 4948 generic.go:334] "Generic (PLEG): container finished" podID="00f4b4c1-af36-4ccd-a3d1-dd6df1186338" containerID="5e9a0379202da2a09189213d51042628595d46ea2ac1801d0bcbd926088826d7" exitCode=0 Feb 20 08:09:06 crc kubenswrapper[4948]: I0220 08:09:06.478016 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n8d7m" event={"ID":"00f4b4c1-af36-4ccd-a3d1-dd6df1186338","Type":"ContainerDied","Data":"5e9a0379202da2a09189213d51042628595d46ea2ac1801d0bcbd926088826d7"} Feb 20 08:09:06 crc kubenswrapper[4948]: I0220 08:09:06.483444 4948 generic.go:334] "Generic (PLEG): container finished" podID="c53ddd33-0f9f-4794-b346-7d48a6c09c9b" containerID="bfea6ab4eabff6af6e0a3b30d304544693e06a2189af921dbd30d2b97a616f83" exitCode=0 Feb 20 08:09:06 crc kubenswrapper[4948]: I0220 08:09:06.483509 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9z42" event={"ID":"c53ddd33-0f9f-4794-b346-7d48a6c09c9b","Type":"ContainerDied","Data":"bfea6ab4eabff6af6e0a3b30d304544693e06a2189af921dbd30d2b97a616f83"} Feb 20 08:09:06 crc kubenswrapper[4948]: I0220 08:09:06.486661 4948 generic.go:334] "Generic (PLEG): container finished" podID="000032f5-8df9-4879-8a8c-7ca06deb0862" containerID="b5372451792db1d5ea16e394e3175c7fb91527d66c015b609f73baf5c0f560d3" exitCode=0 Feb 20 08:09:06 crc kubenswrapper[4948]: I0220 08:09:06.487574 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sbf5n" event={"ID":"000032f5-8df9-4879-8a8c-7ca06deb0862","Type":"ContainerDied","Data":"b5372451792db1d5ea16e394e3175c7fb91527d66c015b609f73baf5c0f560d3"} Feb 20 08:09:06 crc kubenswrapper[4948]: E0220 08:09:06.488913 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-frznz" podUID="dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" Feb 20 08:09:06 crc kubenswrapper[4948]: E0220 08:09:06.491362 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-67ptx" podUID="b8150552-4914-4e92-b041-56600ed6d2f5" Feb 20 08:09:06 crc kubenswrapper[4948]: E0220 08:09:06.491426 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-snc6x" podUID="b5ec9802-aa9c-4518-b99a-1673ab75ec50" Feb 20 08:09:06 crc kubenswrapper[4948]: E0220 08:09:06.493012 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-tvwkh" podUID="b6d696cd-f5cf-47e0-af8d-4d20e93bd80d" Feb 20 08:09:07 crc kubenswrapper[4948]: I0220 08:09:07.492808 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9z42" event={"ID":"c53ddd33-0f9f-4794-b346-7d48a6c09c9b","Type":"ContainerStarted","Data":"1c4c2fe97c46c10efdb2a7e794505136dc520212167c78787ccbac94170a83c0"} Feb 20 08:09:07 crc kubenswrapper[4948]: I0220 08:09:07.515165 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c9z42" podStartSLOduration=2.43809356 podStartE2EDuration="34.515103266s" podCreationTimestamp="2026-02-20 08:08:33 +0000 UTC" firstStartedPulling="2026-02-20 08:08:35.076081143 +0000 UTC m=+164.050575963" lastFinishedPulling="2026-02-20 08:09:07.153090809 +0000 UTC m=+196.127585669" observedRunningTime="2026-02-20 08:09:07.512614805 +0000 UTC m=+196.487109635" watchObservedRunningTime="2026-02-20 08:09:07.515103266 +0000 UTC m=+196.489598086" Feb 20 08:09:08 crc kubenswrapper[4948]: I0220 08:09:08.025433 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:09:08 crc kubenswrapper[4948]: I0220 08:09:08.025913 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:09:08 crc kubenswrapper[4948]: I0220 08:09:08.500565 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sbf5n" event={"ID":"000032f5-8df9-4879-8a8c-7ca06deb0862","Type":"ContainerStarted","Data":"91fe00d50fe88b88435954aa020922438c05af72280d83d2b0cbacc4ca3b0595"} Feb 20 08:09:08 crc kubenswrapper[4948]: I0220 08:09:08.505804 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n8d7m" event={"ID":"00f4b4c1-af36-4ccd-a3d1-dd6df1186338","Type":"ContainerStarted","Data":"acfb3ae1070bc662949f97e63b0d95f3a1957ddfef18898cde66fcbe8186d502"} Feb 20 08:09:08 crc kubenswrapper[4948]: I0220 08:09:08.522612 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-sbf5n" podStartSLOduration=3.963252623 podStartE2EDuration="38.522594053s" podCreationTimestamp="2026-02-20 08:08:30 +0000 UTC" firstStartedPulling="2026-02-20 08:08:32.898081291 +0000 UTC m=+161.872576111" lastFinishedPulling="2026-02-20 08:09:07.457422721 +0000 UTC m=+196.431917541" observedRunningTime="2026-02-20 08:09:08.51961278 +0000 UTC m=+197.494107600" watchObservedRunningTime="2026-02-20 08:09:08.522594053 +0000 UTC m=+197.497088873" Feb 20 08:09:08 crc kubenswrapper[4948]: I0220 08:09:08.537279 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-n8d7m" podStartSLOduration=3.028252157 podStartE2EDuration="37.537258553s" podCreationTimestamp="2026-02-20 08:08:31 +0000 UTC" firstStartedPulling="2026-02-20 08:08:32.909278346 +0000 UTC m=+161.883773166" lastFinishedPulling="2026-02-20 08:09:07.418284742 +0000 UTC m=+196.392779562" observedRunningTime="2026-02-20 08:09:08.53548235 +0000 UTC m=+197.509977170" watchObservedRunningTime="2026-02-20 08:09:08.537258553 +0000 UTC m=+197.511753373" Feb 20 08:09:10 crc kubenswrapper[4948]: I0220 08:09:10.008776 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 20 08:09:10 crc kubenswrapper[4948]: E0220 08:09:10.011557 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b189bd1e-6355-433c-972d-2e27f98fe153" containerName="pruner" Feb 20 08:09:10 crc kubenswrapper[4948]: I0220 08:09:10.011604 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b189bd1e-6355-433c-972d-2e27f98fe153" containerName="pruner" Feb 20 08:09:10 crc kubenswrapper[4948]: E0220 08:09:10.011637 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0" containerName="pruner" Feb 20 08:09:10 crc kubenswrapper[4948]: I0220 08:09:10.011655 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0" containerName="pruner" Feb 20 08:09:10 crc kubenswrapper[4948]: I0220 08:09:10.011856 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="b189bd1e-6355-433c-972d-2e27f98fe153" containerName="pruner" Feb 20 08:09:10 crc kubenswrapper[4948]: I0220 08:09:10.011895 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0" containerName="pruner" Feb 20 08:09:10 crc kubenswrapper[4948]: I0220 08:09:10.012623 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 08:09:10 crc kubenswrapper[4948]: I0220 08:09:10.015953 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 20 08:09:10 crc kubenswrapper[4948]: I0220 08:09:10.017164 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 20 08:09:10 crc kubenswrapper[4948]: I0220 08:09:10.017807 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 20 08:09:11 crc kubenswrapper[4948]: I0220 08:09:11.204122 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb62d1ab-0c70-44c4-a350-ae1c566dbe89-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bb62d1ab-0c70-44c4-a350-ae1c566dbe89\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 08:09:11 crc kubenswrapper[4948]: I0220 08:09:11.204205 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb62d1ab-0c70-44c4-a350-ae1c566dbe89-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bb62d1ab-0c70-44c4-a350-ae1c566dbe89\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 08:09:11 crc kubenswrapper[4948]: I0220 08:09:11.306242 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb62d1ab-0c70-44c4-a350-ae1c566dbe89-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bb62d1ab-0c70-44c4-a350-ae1c566dbe89\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 08:09:11 crc kubenswrapper[4948]: I0220 08:09:11.306415 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb62d1ab-0c70-44c4-a350-ae1c566dbe89-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bb62d1ab-0c70-44c4-a350-ae1c566dbe89\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 08:09:11 crc kubenswrapper[4948]: I0220 08:09:11.306429 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb62d1ab-0c70-44c4-a350-ae1c566dbe89-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bb62d1ab-0c70-44c4-a350-ae1c566dbe89\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 08:09:11 crc kubenswrapper[4948]: I0220 08:09:11.316743 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-sbf5n" Feb 20 08:09:11 crc kubenswrapper[4948]: I0220 08:09:11.316806 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-sbf5n" Feb 20 08:09:11 crc kubenswrapper[4948]: I0220 08:09:11.333032 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb62d1ab-0c70-44c4-a350-ae1c566dbe89-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bb62d1ab-0c70-44c4-a350-ae1c566dbe89\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 08:09:11 crc kubenswrapper[4948]: I0220 08:09:11.424579 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-n8d7m" Feb 20 08:09:11 crc kubenswrapper[4948]: I0220 08:09:11.424642 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-n8d7m" Feb 20 08:09:11 crc kubenswrapper[4948]: I0220 08:09:11.434562 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 08:09:11 crc kubenswrapper[4948]: I0220 08:09:11.529044 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-n8d7m" Feb 20 08:09:11 crc kubenswrapper[4948]: I0220 08:09:11.530852 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-sbf5n" Feb 20 08:09:11 crc kubenswrapper[4948]: I0220 08:09:11.677082 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 20 08:09:11 crc kubenswrapper[4948]: I0220 08:09:11.995453 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sjnxm"] Feb 20 08:09:12 crc kubenswrapper[4948]: I0220 08:09:12.136043 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"bb62d1ab-0c70-44c4-a350-ae1c566dbe89","Type":"ContainerStarted","Data":"b223cee57e1c4f9b3955bdbbeb144ea5ace687c6ba792bf62c3493ca138014b5"} Feb 20 08:09:12 crc kubenswrapper[4948]: I0220 08:09:12.136236 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"bb62d1ab-0c70-44c4-a350-ae1c566dbe89","Type":"ContainerStarted","Data":"a9c38be594b02cd7ae1614b1e36587a85d7f59afa56ac8c0eed109f8018a0ec6"} Feb 20 08:09:12 crc kubenswrapper[4948]: I0220 08:09:12.164328 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=3.164307281 podStartE2EDuration="3.164307281s" podCreationTimestamp="2026-02-20 08:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:09:12.160409876 +0000 UTC m=+201.134904696" watchObservedRunningTime="2026-02-20 08:09:12.164307281 +0000 UTC m=+201.138802101" Feb 20 08:09:12 crc kubenswrapper[4948]: I0220 08:09:12.197921 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-n8d7m" Feb 20 08:09:12 crc kubenswrapper[4948]: I0220 08:09:12.202751 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-sbf5n" Feb 20 08:09:13 crc kubenswrapper[4948]: I0220 08:09:13.142125 4948 generic.go:334] "Generic (PLEG): container finished" podID="bb62d1ab-0c70-44c4-a350-ae1c566dbe89" containerID="b223cee57e1c4f9b3955bdbbeb144ea5ace687c6ba792bf62c3493ca138014b5" exitCode=0 Feb 20 08:09:13 crc kubenswrapper[4948]: I0220 08:09:13.142207 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"bb62d1ab-0c70-44c4-a350-ae1c566dbe89","Type":"ContainerDied","Data":"b223cee57e1c4f9b3955bdbbeb144ea5ace687c6ba792bf62c3493ca138014b5"} Feb 20 08:09:13 crc kubenswrapper[4948]: I0220 08:09:13.144444 4948 generic.go:334] "Generic (PLEG): container finished" podID="f75a50cf-7dae-420b-a00d-671ede3cb6f7" containerID="ed76ff8cc753dd2cd048a3406074573b902071e966e21613ec37f782fb620d37" exitCode=0 Feb 20 08:09:13 crc kubenswrapper[4948]: I0220 08:09:13.144532 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nrhl6" event={"ID":"f75a50cf-7dae-420b-a00d-671ede3cb6f7","Type":"ContainerDied","Data":"ed76ff8cc753dd2cd048a3406074573b902071e966e21613ec37f782fb620d37"} Feb 20 08:09:13 crc kubenswrapper[4948]: I0220 08:09:13.422351 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c9z42" Feb 20 08:09:13 crc kubenswrapper[4948]: I0220 08:09:13.422526 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c9z42" Feb 20 08:09:13 crc kubenswrapper[4948]: I0220 08:09:13.467643 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c9z42" Feb 20 08:09:14 crc kubenswrapper[4948]: I0220 08:09:14.152140 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nrhl6" event={"ID":"f75a50cf-7dae-420b-a00d-671ede3cb6f7","Type":"ContainerStarted","Data":"51c14bb6285644fd84e90599149f7f330a4395b58ff4509817fbdce179d6470b"} Feb 20 08:09:14 crc kubenswrapper[4948]: I0220 08:09:14.171473 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nrhl6" podStartSLOduration=2.258996389 podStartE2EDuration="44.171444284s" podCreationTimestamp="2026-02-20 08:08:30 +0000 UTC" firstStartedPulling="2026-02-20 08:08:31.893208258 +0000 UTC m=+160.867703078" lastFinishedPulling="2026-02-20 08:09:13.805656113 +0000 UTC m=+202.780150973" observedRunningTime="2026-02-20 08:09:14.169593548 +0000 UTC m=+203.144088378" watchObservedRunningTime="2026-02-20 08:09:14.171444284 +0000 UTC m=+203.145939124" Feb 20 08:09:14 crc kubenswrapper[4948]: I0220 08:09:14.207729 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c9z42" Feb 20 08:09:14 crc kubenswrapper[4948]: I0220 08:09:14.398737 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 08:09:14 crc kubenswrapper[4948]: I0220 08:09:14.570593 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb62d1ab-0c70-44c4-a350-ae1c566dbe89-kubelet-dir\") pod \"bb62d1ab-0c70-44c4-a350-ae1c566dbe89\" (UID: \"bb62d1ab-0c70-44c4-a350-ae1c566dbe89\") " Feb 20 08:09:14 crc kubenswrapper[4948]: I0220 08:09:14.570750 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bb62d1ab-0c70-44c4-a350-ae1c566dbe89-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "bb62d1ab-0c70-44c4-a350-ae1c566dbe89" (UID: "bb62d1ab-0c70-44c4-a350-ae1c566dbe89"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:09:14 crc kubenswrapper[4948]: I0220 08:09:14.570780 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb62d1ab-0c70-44c4-a350-ae1c566dbe89-kube-api-access\") pod \"bb62d1ab-0c70-44c4-a350-ae1c566dbe89\" (UID: \"bb62d1ab-0c70-44c4-a350-ae1c566dbe89\") " Feb 20 08:09:14 crc kubenswrapper[4948]: I0220 08:09:14.571847 4948 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb62d1ab-0c70-44c4-a350-ae1c566dbe89-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:14 crc kubenswrapper[4948]: I0220 08:09:14.582137 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb62d1ab-0c70-44c4-a350-ae1c566dbe89-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "bb62d1ab-0c70-44c4-a350-ae1c566dbe89" (UID: "bb62d1ab-0c70-44c4-a350-ae1c566dbe89"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:09:14 crc kubenswrapper[4948]: I0220 08:09:14.672877 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb62d1ab-0c70-44c4-a350-ae1c566dbe89-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:14 crc kubenswrapper[4948]: I0220 08:09:14.713016 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sbf5n"] Feb 20 08:09:14 crc kubenswrapper[4948]: I0220 08:09:14.713241 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-sbf5n" podUID="000032f5-8df9-4879-8a8c-7ca06deb0862" containerName="registry-server" containerID="cri-o://91fe00d50fe88b88435954aa020922438c05af72280d83d2b0cbacc4ca3b0595" gracePeriod=2 Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.061426 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sbf5n" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.158735 4948 generic.go:334] "Generic (PLEG): container finished" podID="000032f5-8df9-4879-8a8c-7ca06deb0862" containerID="91fe00d50fe88b88435954aa020922438c05af72280d83d2b0cbacc4ca3b0595" exitCode=0 Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.158802 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sbf5n" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.158808 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sbf5n" event={"ID":"000032f5-8df9-4879-8a8c-7ca06deb0862","Type":"ContainerDied","Data":"91fe00d50fe88b88435954aa020922438c05af72280d83d2b0cbacc4ca3b0595"} Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.158842 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sbf5n" event={"ID":"000032f5-8df9-4879-8a8c-7ca06deb0862","Type":"ContainerDied","Data":"6682a9f05af301b707c145d9ce47468c3451f6e6fc6b07c7b78c99356b765594"} Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.158860 4948 scope.go:117] "RemoveContainer" containerID="91fe00d50fe88b88435954aa020922438c05af72280d83d2b0cbacc4ca3b0595" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.160494 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.160522 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"bb62d1ab-0c70-44c4-a350-ae1c566dbe89","Type":"ContainerDied","Data":"a9c38be594b02cd7ae1614b1e36587a85d7f59afa56ac8c0eed109f8018a0ec6"} Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.160563 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9c38be594b02cd7ae1614b1e36587a85d7f59afa56ac8c0eed109f8018a0ec6" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.177828 4948 scope.go:117] "RemoveContainer" containerID="b5372451792db1d5ea16e394e3175c7fb91527d66c015b609f73baf5c0f560d3" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.179431 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqxvg\" (UniqueName: \"kubernetes.io/projected/000032f5-8df9-4879-8a8c-7ca06deb0862-kube-api-access-dqxvg\") pod \"000032f5-8df9-4879-8a8c-7ca06deb0862\" (UID: \"000032f5-8df9-4879-8a8c-7ca06deb0862\") " Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.179522 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/000032f5-8df9-4879-8a8c-7ca06deb0862-utilities\") pod \"000032f5-8df9-4879-8a8c-7ca06deb0862\" (UID: \"000032f5-8df9-4879-8a8c-7ca06deb0862\") " Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.179863 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/000032f5-8df9-4879-8a8c-7ca06deb0862-catalog-content\") pod \"000032f5-8df9-4879-8a8c-7ca06deb0862\" (UID: \"000032f5-8df9-4879-8a8c-7ca06deb0862\") " Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.180326 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/000032f5-8df9-4879-8a8c-7ca06deb0862-utilities" (OuterVolumeSpecName: "utilities") pod "000032f5-8df9-4879-8a8c-7ca06deb0862" (UID: "000032f5-8df9-4879-8a8c-7ca06deb0862"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.184176 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/000032f5-8df9-4879-8a8c-7ca06deb0862-kube-api-access-dqxvg" (OuterVolumeSpecName: "kube-api-access-dqxvg") pod "000032f5-8df9-4879-8a8c-7ca06deb0862" (UID: "000032f5-8df9-4879-8a8c-7ca06deb0862"). InnerVolumeSpecName "kube-api-access-dqxvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.224282 4948 scope.go:117] "RemoveContainer" containerID="b939d98be77246f833198ae82a9cce74eaf753fb0b14e37eb3663fc9e452cd93" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.240957 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/000032f5-8df9-4879-8a8c-7ca06deb0862-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "000032f5-8df9-4879-8a8c-7ca06deb0862" (UID: "000032f5-8df9-4879-8a8c-7ca06deb0862"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.255638 4948 scope.go:117] "RemoveContainer" containerID="91fe00d50fe88b88435954aa020922438c05af72280d83d2b0cbacc4ca3b0595" Feb 20 08:09:15 crc kubenswrapper[4948]: E0220 08:09:15.256135 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91fe00d50fe88b88435954aa020922438c05af72280d83d2b0cbacc4ca3b0595\": container with ID starting with 91fe00d50fe88b88435954aa020922438c05af72280d83d2b0cbacc4ca3b0595 not found: ID does not exist" containerID="91fe00d50fe88b88435954aa020922438c05af72280d83d2b0cbacc4ca3b0595" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.256183 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91fe00d50fe88b88435954aa020922438c05af72280d83d2b0cbacc4ca3b0595"} err="failed to get container status \"91fe00d50fe88b88435954aa020922438c05af72280d83d2b0cbacc4ca3b0595\": rpc error: code = NotFound desc = could not find container \"91fe00d50fe88b88435954aa020922438c05af72280d83d2b0cbacc4ca3b0595\": container with ID starting with 91fe00d50fe88b88435954aa020922438c05af72280d83d2b0cbacc4ca3b0595 not found: ID does not exist" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.256242 4948 scope.go:117] "RemoveContainer" containerID="b5372451792db1d5ea16e394e3175c7fb91527d66c015b609f73baf5c0f560d3" Feb 20 08:09:15 crc kubenswrapper[4948]: E0220 08:09:15.256750 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5372451792db1d5ea16e394e3175c7fb91527d66c015b609f73baf5c0f560d3\": container with ID starting with b5372451792db1d5ea16e394e3175c7fb91527d66c015b609f73baf5c0f560d3 not found: ID does not exist" containerID="b5372451792db1d5ea16e394e3175c7fb91527d66c015b609f73baf5c0f560d3" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.256771 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5372451792db1d5ea16e394e3175c7fb91527d66c015b609f73baf5c0f560d3"} err="failed to get container status \"b5372451792db1d5ea16e394e3175c7fb91527d66c015b609f73baf5c0f560d3\": rpc error: code = NotFound desc = could not find container \"b5372451792db1d5ea16e394e3175c7fb91527d66c015b609f73baf5c0f560d3\": container with ID starting with b5372451792db1d5ea16e394e3175c7fb91527d66c015b609f73baf5c0f560d3 not found: ID does not exist" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.256785 4948 scope.go:117] "RemoveContainer" containerID="b939d98be77246f833198ae82a9cce74eaf753fb0b14e37eb3663fc9e452cd93" Feb 20 08:09:15 crc kubenswrapper[4948]: E0220 08:09:15.256989 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b939d98be77246f833198ae82a9cce74eaf753fb0b14e37eb3663fc9e452cd93\": container with ID starting with b939d98be77246f833198ae82a9cce74eaf753fb0b14e37eb3663fc9e452cd93 not found: ID does not exist" containerID="b939d98be77246f833198ae82a9cce74eaf753fb0b14e37eb3663fc9e452cd93" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.257008 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b939d98be77246f833198ae82a9cce74eaf753fb0b14e37eb3663fc9e452cd93"} err="failed to get container status \"b939d98be77246f833198ae82a9cce74eaf753fb0b14e37eb3663fc9e452cd93\": rpc error: code = NotFound desc = could not find container \"b939d98be77246f833198ae82a9cce74eaf753fb0b14e37eb3663fc9e452cd93\": container with ID starting with b939d98be77246f833198ae82a9cce74eaf753fb0b14e37eb3663fc9e452cd93 not found: ID does not exist" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.281099 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqxvg\" (UniqueName: \"kubernetes.io/projected/000032f5-8df9-4879-8a8c-7ca06deb0862-kube-api-access-dqxvg\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.281136 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/000032f5-8df9-4879-8a8c-7ca06deb0862-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.281145 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/000032f5-8df9-4879-8a8c-7ca06deb0862-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.487855 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sbf5n"] Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.492459 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-sbf5n"] Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.759597 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="000032f5-8df9-4879-8a8c-7ca06deb0862" path="/var/lib/kubelet/pods/000032f5-8df9-4879-8a8c-7ca06deb0862/volumes" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.002100 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 20 08:09:16 crc kubenswrapper[4948]: E0220 08:09:16.002392 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="000032f5-8df9-4879-8a8c-7ca06deb0862" containerName="extract-content" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.002418 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="000032f5-8df9-4879-8a8c-7ca06deb0862" containerName="extract-content" Feb 20 08:09:16 crc kubenswrapper[4948]: E0220 08:09:16.002434 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="000032f5-8df9-4879-8a8c-7ca06deb0862" containerName="extract-utilities" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.002445 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="000032f5-8df9-4879-8a8c-7ca06deb0862" containerName="extract-utilities" Feb 20 08:09:16 crc kubenswrapper[4948]: E0220 08:09:16.002459 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb62d1ab-0c70-44c4-a350-ae1c566dbe89" containerName="pruner" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.002471 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb62d1ab-0c70-44c4-a350-ae1c566dbe89" containerName="pruner" Feb 20 08:09:16 crc kubenswrapper[4948]: E0220 08:09:16.002489 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="000032f5-8df9-4879-8a8c-7ca06deb0862" containerName="registry-server" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.002499 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="000032f5-8df9-4879-8a8c-7ca06deb0862" containerName="registry-server" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.002641 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb62d1ab-0c70-44c4-a350-ae1c566dbe89" containerName="pruner" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.002658 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="000032f5-8df9-4879-8a8c-7ca06deb0862" containerName="registry-server" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.003145 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.005003 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.005150 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.013241 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.191461 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.191519 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-var-lock\") pod \"installer-9-crc\" (UID: \"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.191552 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-kube-api-access\") pod \"installer-9-crc\" (UID: \"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.292853 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.292895 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.292965 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-var-lock\") pod \"installer-9-crc\" (UID: \"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.293054 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-kube-api-access\") pod \"installer-9-crc\" (UID: \"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.293158 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-var-lock\") pod \"installer-9-crc\" (UID: \"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.312698 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-kube-api-access\") pod \"installer-9-crc\" (UID: \"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.323337 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.781292 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 20 08:09:16 crc kubenswrapper[4948]: W0220 08:09:16.790601 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-poda0fc48a9_8f2e_4a58_9ec2_0ab67f6cedea.slice/crio-501f50c7bcee3c3ff4d799d7aadad96a2760427edf1eb9577fa78b654115cf07 WatchSource:0}: Error finding container 501f50c7bcee3c3ff4d799d7aadad96a2760427edf1eb9577fa78b654115cf07: Status 404 returned error can't find the container with id 501f50c7bcee3c3ff4d799d7aadad96a2760427edf1eb9577fa78b654115cf07 Feb 20 08:09:17 crc kubenswrapper[4948]: I0220 08:09:17.173211 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea","Type":"ContainerStarted","Data":"501f50c7bcee3c3ff4d799d7aadad96a2760427edf1eb9577fa78b654115cf07"} Feb 20 08:09:18 crc kubenswrapper[4948]: I0220 08:09:18.179885 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea","Type":"ContainerStarted","Data":"308dd2b5632b47b59d9d195703329e9e330103427c0967b20ee97ba81766b1b3"} Feb 20 08:09:18 crc kubenswrapper[4948]: I0220 08:09:18.198821 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=3.198793508 podStartE2EDuration="3.198793508s" podCreationTimestamp="2026-02-20 08:09:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:09:18.197170358 +0000 UTC m=+207.171665218" watchObservedRunningTime="2026-02-20 08:09:18.198793508 +0000 UTC m=+207.173288368" Feb 20 08:09:20 crc kubenswrapper[4948]: I0220 08:09:20.772494 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nrhl6" Feb 20 08:09:20 crc kubenswrapper[4948]: I0220 08:09:20.773606 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nrhl6" Feb 20 08:09:20 crc kubenswrapper[4948]: I0220 08:09:20.813608 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nrhl6" Feb 20 08:09:21 crc kubenswrapper[4948]: I0220 08:09:21.313391 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nrhl6" Feb 20 08:09:27 crc kubenswrapper[4948]: I0220 08:09:27.315934 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-snc6x" event={"ID":"b5ec9802-aa9c-4518-b99a-1673ab75ec50","Type":"ContainerStarted","Data":"e970f5a271f9d8c2873e14eacbf1c841136223bb57e91d7713eb158290760f1e"} Feb 20 08:09:27 crc kubenswrapper[4948]: I0220 08:09:27.318366 4948 generic.go:334] "Generic (PLEG): container finished" podID="b6d696cd-f5cf-47e0-af8d-4d20e93bd80d" containerID="adec323384ee4b89adc829f6e8cec6c7df59b4806363aa0408583b7df2842893" exitCode=0 Feb 20 08:09:27 crc kubenswrapper[4948]: I0220 08:09:27.318416 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tvwkh" event={"ID":"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d","Type":"ContainerDied","Data":"adec323384ee4b89adc829f6e8cec6c7df59b4806363aa0408583b7df2842893"} Feb 20 08:09:28 crc kubenswrapper[4948]: I0220 08:09:28.326270 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frznz" event={"ID":"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9","Type":"ContainerStarted","Data":"27b14a9abd1358ef732430763a33e8727bba43c949149d1b6052fc586fb0bd00"} Feb 20 08:09:28 crc kubenswrapper[4948]: I0220 08:09:28.328875 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-67ptx" event={"ID":"b8150552-4914-4e92-b041-56600ed6d2f5","Type":"ContainerStarted","Data":"e9c973e2d8ac9c5ce88b599645da506dbb8f6dbaabbec1d3e5fca44ff09a3d5c"} Feb 20 08:09:28 crc kubenswrapper[4948]: I0220 08:09:28.330782 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-snc6x" event={"ID":"b5ec9802-aa9c-4518-b99a-1673ab75ec50","Type":"ContainerDied","Data":"e970f5a271f9d8c2873e14eacbf1c841136223bb57e91d7713eb158290760f1e"} Feb 20 08:09:28 crc kubenswrapper[4948]: I0220 08:09:28.330681 4948 generic.go:334] "Generic (PLEG): container finished" podID="b5ec9802-aa9c-4518-b99a-1673ab75ec50" containerID="e970f5a271f9d8c2873e14eacbf1c841136223bb57e91d7713eb158290760f1e" exitCode=0 Feb 20 08:09:29 crc kubenswrapper[4948]: I0220 08:09:29.339519 4948 generic.go:334] "Generic (PLEG): container finished" podID="dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" containerID="27b14a9abd1358ef732430763a33e8727bba43c949149d1b6052fc586fb0bd00" exitCode=0 Feb 20 08:09:29 crc kubenswrapper[4948]: I0220 08:09:29.339619 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frznz" event={"ID":"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9","Type":"ContainerDied","Data":"27b14a9abd1358ef732430763a33e8727bba43c949149d1b6052fc586fb0bd00"} Feb 20 08:09:29 crc kubenswrapper[4948]: I0220 08:09:29.343571 4948 generic.go:334] "Generic (PLEG): container finished" podID="b8150552-4914-4e92-b041-56600ed6d2f5" containerID="e9c973e2d8ac9c5ce88b599645da506dbb8f6dbaabbec1d3e5fca44ff09a3d5c" exitCode=0 Feb 20 08:09:29 crc kubenswrapper[4948]: I0220 08:09:29.343657 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-67ptx" event={"ID":"b8150552-4914-4e92-b041-56600ed6d2f5","Type":"ContainerDied","Data":"e9c973e2d8ac9c5ce88b599645da506dbb8f6dbaabbec1d3e5fca44ff09a3d5c"} Feb 20 08:09:29 crc kubenswrapper[4948]: I0220 08:09:29.346273 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-snc6x" event={"ID":"b5ec9802-aa9c-4518-b99a-1673ab75ec50","Type":"ContainerStarted","Data":"59bc680a37628dd6760eae32b4fd77ee425353925297c0763e3ca5bf6107c49e"} Feb 20 08:09:29 crc kubenswrapper[4948]: I0220 08:09:29.348992 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tvwkh" event={"ID":"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d","Type":"ContainerStarted","Data":"3eb57d35e2ca5917ed63ba672507ebdc1b840d438dcdc8930f619e051d58336d"} Feb 20 08:09:29 crc kubenswrapper[4948]: I0220 08:09:29.427344 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-snc6x" podStartSLOduration=2.684655272 podStartE2EDuration="55.42730509s" podCreationTimestamp="2026-02-20 08:08:34 +0000 UTC" firstStartedPulling="2026-02-20 08:08:36.067381954 +0000 UTC m=+165.041876774" lastFinishedPulling="2026-02-20 08:09:28.810031772 +0000 UTC m=+217.784526592" observedRunningTime="2026-02-20 08:09:29.398590686 +0000 UTC m=+218.373085526" watchObservedRunningTime="2026-02-20 08:09:29.42730509 +0000 UTC m=+218.401799930" Feb 20 08:09:29 crc kubenswrapper[4948]: I0220 08:09:29.451205 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tvwkh" podStartSLOduration=3.612786143 podStartE2EDuration="58.451179756s" podCreationTimestamp="2026-02-20 08:08:31 +0000 UTC" firstStartedPulling="2026-02-20 08:08:33.957213385 +0000 UTC m=+162.931708205" lastFinishedPulling="2026-02-20 08:09:28.795606998 +0000 UTC m=+217.770101818" observedRunningTime="2026-02-20 08:09:29.44893173 +0000 UTC m=+218.423426550" watchObservedRunningTime="2026-02-20 08:09:29.451179756 +0000 UTC m=+218.425674576" Feb 20 08:09:31 crc kubenswrapper[4948]: I0220 08:09:31.362526 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frznz" event={"ID":"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9","Type":"ContainerStarted","Data":"fb097a14ed0ab7d5af6abab587a1c01e68baf8b43ccad645f65796167b0c90bd"} Feb 20 08:09:31 crc kubenswrapper[4948]: I0220 08:09:31.365223 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-67ptx" event={"ID":"b8150552-4914-4e92-b041-56600ed6d2f5","Type":"ContainerStarted","Data":"3ece74a70c863ce99b2f2f650f2d87de55dd27277341567aa5ed99e67f8a4dc8"} Feb 20 08:09:31 crc kubenswrapper[4948]: I0220 08:09:31.407114 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-frznz" podStartSLOduration=3.17650005 podStartE2EDuration="58.407089222s" podCreationTimestamp="2026-02-20 08:08:33 +0000 UTC" firstStartedPulling="2026-02-20 08:08:35.075987961 +0000 UTC m=+164.050482781" lastFinishedPulling="2026-02-20 08:09:30.306577123 +0000 UTC m=+219.281071953" observedRunningTime="2026-02-20 08:09:31.385935723 +0000 UTC m=+220.360430543" watchObservedRunningTime="2026-02-20 08:09:31.407089222 +0000 UTC m=+220.381584052" Feb 20 08:09:31 crc kubenswrapper[4948]: I0220 08:09:31.408878 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-67ptx" podStartSLOduration=2.779394732 podStartE2EDuration="58.408867796s" podCreationTimestamp="2026-02-20 08:08:33 +0000 UTC" firstStartedPulling="2026-02-20 08:08:35.0759414 +0000 UTC m=+164.050436220" lastFinishedPulling="2026-02-20 08:09:30.705414424 +0000 UTC m=+219.679909284" observedRunningTime="2026-02-20 08:09:31.404619171 +0000 UTC m=+220.379114001" watchObservedRunningTime="2026-02-20 08:09:31.408867796 +0000 UTC m=+220.383362636" Feb 20 08:09:31 crc kubenswrapper[4948]: I0220 08:09:31.631557 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tvwkh" Feb 20 08:09:31 crc kubenswrapper[4948]: I0220 08:09:31.631632 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tvwkh" Feb 20 08:09:31 crc kubenswrapper[4948]: I0220 08:09:31.688041 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tvwkh" Feb 20 08:09:33 crc kubenswrapper[4948]: I0220 08:09:33.795831 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-67ptx" Feb 20 08:09:33 crc kubenswrapper[4948]: I0220 08:09:33.796950 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-67ptx" Feb 20 08:09:33 crc kubenswrapper[4948]: I0220 08:09:33.860217 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-67ptx" Feb 20 08:09:34 crc kubenswrapper[4948]: I0220 08:09:34.216224 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-frznz" Feb 20 08:09:34 crc kubenswrapper[4948]: I0220 08:09:34.216278 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-frznz" Feb 20 08:09:34 crc kubenswrapper[4948]: I0220 08:09:34.467037 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-snc6x" Feb 20 08:09:34 crc kubenswrapper[4948]: I0220 08:09:34.467355 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-snc6x" Feb 20 08:09:35 crc kubenswrapper[4948]: I0220 08:09:35.257044 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-frznz" podUID="dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" containerName="registry-server" probeResult="failure" output=< Feb 20 08:09:35 crc kubenswrapper[4948]: timeout: failed to connect service ":50051" within 1s Feb 20 08:09:35 crc kubenswrapper[4948]: > Feb 20 08:09:35 crc kubenswrapper[4948]: I0220 08:09:35.505444 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-snc6x" podUID="b5ec9802-aa9c-4518-b99a-1673ab75ec50" containerName="registry-server" probeResult="failure" output=< Feb 20 08:09:35 crc kubenswrapper[4948]: timeout: failed to connect service ":50051" within 1s Feb 20 08:09:35 crc kubenswrapper[4948]: > Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.030030 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" podUID="df73228e-df1f-4e9c-aafe-493bb743f98a" containerName="oauth-openshift" containerID="cri-o://f0e4fad62991d25dea391bf5b9e653af0d194dd88ba8a814b4a27d229ce2ef30" gracePeriod=15 Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.380893 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.396466 4948 generic.go:334] "Generic (PLEG): container finished" podID="df73228e-df1f-4e9c-aafe-493bb743f98a" containerID="f0e4fad62991d25dea391bf5b9e653af0d194dd88ba8a814b4a27d229ce2ef30" exitCode=0 Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.396510 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" event={"ID":"df73228e-df1f-4e9c-aafe-493bb743f98a","Type":"ContainerDied","Data":"f0e4fad62991d25dea391bf5b9e653af0d194dd88ba8a814b4a27d229ce2ef30"} Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.396533 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" event={"ID":"df73228e-df1f-4e9c-aafe-493bb743f98a","Type":"ContainerDied","Data":"76dafdb5407bf70db2d0cf3c70d0d7427c429ce3ab8c2b2e1a871cb56382ee9a"} Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.396549 4948 scope.go:117] "RemoveContainer" containerID="f0e4fad62991d25dea391bf5b9e653af0d194dd88ba8a814b4a27d229ce2ef30" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.396669 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.417701 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-d96b794dc-r7qn6"] Feb 20 08:09:37 crc kubenswrapper[4948]: E0220 08:09:37.418313 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df73228e-df1f-4e9c-aafe-493bb743f98a" containerName="oauth-openshift" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.418326 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="df73228e-df1f-4e9c-aafe-493bb743f98a" containerName="oauth-openshift" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.418452 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="df73228e-df1f-4e9c-aafe-493bb743f98a" containerName="oauth-openshift" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.418909 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.421388 4948 scope.go:117] "RemoveContainer" containerID="f0e4fad62991d25dea391bf5b9e653af0d194dd88ba8a814b4a27d229ce2ef30" Feb 20 08:09:37 crc kubenswrapper[4948]: E0220 08:09:37.421847 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0e4fad62991d25dea391bf5b9e653af0d194dd88ba8a814b4a27d229ce2ef30\": container with ID starting with f0e4fad62991d25dea391bf5b9e653af0d194dd88ba8a814b4a27d229ce2ef30 not found: ID does not exist" containerID="f0e4fad62991d25dea391bf5b9e653af0d194dd88ba8a814b4a27d229ce2ef30" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.421871 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0e4fad62991d25dea391bf5b9e653af0d194dd88ba8a814b4a27d229ce2ef30"} err="failed to get container status \"f0e4fad62991d25dea391bf5b9e653af0d194dd88ba8a814b4a27d229ce2ef30\": rpc error: code = NotFound desc = could not find container \"f0e4fad62991d25dea391bf5b9e653af0d194dd88ba8a814b4a27d229ce2ef30\": container with ID starting with f0e4fad62991d25dea391bf5b9e653af0d194dd88ba8a814b4a27d229ce2ef30 not found: ID does not exist" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.445480 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-ocp-branding-template\") pod \"df73228e-df1f-4e9c-aafe-493bb743f98a\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.446004 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-service-ca\") pod \"df73228e-df1f-4e9c-aafe-493bb743f98a\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.446144 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2l9z\" (UniqueName: \"kubernetes.io/projected/df73228e-df1f-4e9c-aafe-493bb743f98a-kube-api-access-q2l9z\") pod \"df73228e-df1f-4e9c-aafe-493bb743f98a\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.446430 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-service-ca\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.446459 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-user-template-login\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.446479 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5f01646f-0fb2-40ee-81aa-77f177add987-audit-dir\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.446665 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.446690 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.446828 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-session\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.446859 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-serving-cert\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.447038 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-router-certs\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.447058 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czs2d\" (UniqueName: \"kubernetes.io/projected/5f01646f-0fb2-40ee-81aa-77f177add987-kube-api-access-czs2d\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.447078 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5f01646f-0fb2-40ee-81aa-77f177add987-audit-policies\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.447276 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.447300 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.447320 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-user-template-error\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.447448 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-cliconfig\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.448075 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "df73228e-df1f-4e9c-aafe-493bb743f98a" (UID: "df73228e-df1f-4e9c-aafe-493bb743f98a"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.453456 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "df73228e-df1f-4e9c-aafe-493bb743f98a" (UID: "df73228e-df1f-4e9c-aafe-493bb743f98a"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.454062 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df73228e-df1f-4e9c-aafe-493bb743f98a-kube-api-access-q2l9z" (OuterVolumeSpecName: "kube-api-access-q2l9z") pod "df73228e-df1f-4e9c-aafe-493bb743f98a" (UID: "df73228e-df1f-4e9c-aafe-493bb743f98a"). InnerVolumeSpecName "kube-api-access-q2l9z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.474543 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-d96b794dc-r7qn6"] Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.548386 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-provider-selection\") pod \"df73228e-df1f-4e9c-aafe-493bb743f98a\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.548507 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-trusted-ca-bundle\") pod \"df73228e-df1f-4e9c-aafe-493bb743f98a\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.548579 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-cliconfig\") pod \"df73228e-df1f-4e9c-aafe-493bb743f98a\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.548624 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-session\") pod \"df73228e-df1f-4e9c-aafe-493bb743f98a\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.548656 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-error\") pod \"df73228e-df1f-4e9c-aafe-493bb743f98a\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.548678 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-login\") pod \"df73228e-df1f-4e9c-aafe-493bb743f98a\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.548715 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-idp-0-file-data\") pod \"df73228e-df1f-4e9c-aafe-493bb743f98a\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.548744 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/df73228e-df1f-4e9c-aafe-493bb743f98a-audit-dir\") pod \"df73228e-df1f-4e9c-aafe-493bb743f98a\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.548770 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-audit-policies\") pod \"df73228e-df1f-4e9c-aafe-493bb743f98a\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.548796 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-router-certs\") pod \"df73228e-df1f-4e9c-aafe-493bb743f98a\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.548816 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-serving-cert\") pod \"df73228e-df1f-4e9c-aafe-493bb743f98a\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.548958 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-serving-cert\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.549007 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-router-certs\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.549036 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czs2d\" (UniqueName: \"kubernetes.io/projected/5f01646f-0fb2-40ee-81aa-77f177add987-kube-api-access-czs2d\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.549061 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5f01646f-0fb2-40ee-81aa-77f177add987-audit-policies\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.549083 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.549103 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.549123 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-user-template-error\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.549149 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-cliconfig\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.549171 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-service-ca\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.549200 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-user-template-login\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.549219 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5f01646f-0fb2-40ee-81aa-77f177add987-audit-dir\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.549241 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.549261 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.549288 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-session\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.549322 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.549334 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2l9z\" (UniqueName: \"kubernetes.io/projected/df73228e-df1f-4e9c-aafe-493bb743f98a-kube-api-access-q2l9z\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.549345 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.549443 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "df73228e-df1f-4e9c-aafe-493bb743f98a" (UID: "df73228e-df1f-4e9c-aafe-493bb743f98a"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.551134 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5f01646f-0fb2-40ee-81aa-77f177add987-audit-policies\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.551225 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5f01646f-0fb2-40ee-81aa-77f177add987-audit-dir\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.551360 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "df73228e-df1f-4e9c-aafe-493bb743f98a" (UID: "df73228e-df1f-4e9c-aafe-493bb743f98a"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.551673 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df73228e-df1f-4e9c-aafe-493bb743f98a-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "df73228e-df1f-4e9c-aafe-493bb743f98a" (UID: "df73228e-df1f-4e9c-aafe-493bb743f98a"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.552204 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-service-ca\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.552752 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "df73228e-df1f-4e9c-aafe-493bb743f98a" (UID: "df73228e-df1f-4e9c-aafe-493bb743f98a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.552854 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-cliconfig\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.554132 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.557606 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "df73228e-df1f-4e9c-aafe-493bb743f98a" (UID: "df73228e-df1f-4e9c-aafe-493bb743f98a"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.559191 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-router-certs\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.559292 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-session\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.559931 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.561162 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.570189 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-serving-cert\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.570296 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-user-template-login\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.570396 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "df73228e-df1f-4e9c-aafe-493bb743f98a" (UID: "df73228e-df1f-4e9c-aafe-493bb743f98a"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.570625 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-user-template-error\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.570791 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "df73228e-df1f-4e9c-aafe-493bb743f98a" (UID: "df73228e-df1f-4e9c-aafe-493bb743f98a"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.570961 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "df73228e-df1f-4e9c-aafe-493bb743f98a" (UID: "df73228e-df1f-4e9c-aafe-493bb743f98a"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.571185 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "df73228e-df1f-4e9c-aafe-493bb743f98a" (UID: "df73228e-df1f-4e9c-aafe-493bb743f98a"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.571535 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "df73228e-df1f-4e9c-aafe-493bb743f98a" (UID: "df73228e-df1f-4e9c-aafe-493bb743f98a"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.571676 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "df73228e-df1f-4e9c-aafe-493bb743f98a" (UID: "df73228e-df1f-4e9c-aafe-493bb743f98a"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.574310 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.574617 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czs2d\" (UniqueName: \"kubernetes.io/projected/5f01646f-0fb2-40ee-81aa-77f177add987-kube-api-access-czs2d\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.651137 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.651198 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.651221 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.651242 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.651262 4948 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/df73228e-df1f-4e9c-aafe-493bb743f98a-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.651280 4948 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.651298 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.651315 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.651335 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.651353 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.651371 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.733358 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.748407 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sjnxm"] Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.750475 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sjnxm"] Feb 20 08:09:38 crc kubenswrapper[4948]: I0220 08:09:38.025672 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:09:38 crc kubenswrapper[4948]: I0220 08:09:38.025770 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:09:38 crc kubenswrapper[4948]: I0220 08:09:38.025839 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:09:38 crc kubenswrapper[4948]: I0220 08:09:38.026960 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59"} pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 08:09:38 crc kubenswrapper[4948]: I0220 08:09:38.027103 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" containerID="cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59" gracePeriod=600 Feb 20 08:09:38 crc kubenswrapper[4948]: I0220 08:09:38.257337 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-d96b794dc-r7qn6"] Feb 20 08:09:38 crc kubenswrapper[4948]: W0220 08:09:38.265512 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f01646f_0fb2_40ee_81aa_77f177add987.slice/crio-0b2fc2bc7a6727af6636f96647e0bcbf84785fef9ee71daef8751c86bc59125b WatchSource:0}: Error finding container 0b2fc2bc7a6727af6636f96647e0bcbf84785fef9ee71daef8751c86bc59125b: Status 404 returned error can't find the container with id 0b2fc2bc7a6727af6636f96647e0bcbf84785fef9ee71daef8751c86bc59125b Feb 20 08:09:38 crc kubenswrapper[4948]: I0220 08:09:38.408412 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" event={"ID":"5f01646f-0fb2-40ee-81aa-77f177add987","Type":"ContainerStarted","Data":"0b2fc2bc7a6727af6636f96647e0bcbf84785fef9ee71daef8751c86bc59125b"} Feb 20 08:09:39 crc kubenswrapper[4948]: I0220 08:09:39.420498 4948 generic.go:334] "Generic (PLEG): container finished" podID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerID="874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59" exitCode=0 Feb 20 08:09:39 crc kubenswrapper[4948]: I0220 08:09:39.420598 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerDied","Data":"874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59"} Feb 20 08:09:39 crc kubenswrapper[4948]: I0220 08:09:39.421145 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"6791802a782456e3c45df951fa55097ce0ea3affd6a9878fc2981511046d1b09"} Feb 20 08:09:39 crc kubenswrapper[4948]: I0220 08:09:39.424378 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" event={"ID":"5f01646f-0fb2-40ee-81aa-77f177add987","Type":"ContainerStarted","Data":"8e95464b61e8b22fef4ef903732e5929f250c789b181c6c0c192edf0fbeb96ba"} Feb 20 08:09:39 crc kubenswrapper[4948]: I0220 08:09:39.427177 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:39 crc kubenswrapper[4948]: I0220 08:09:39.434101 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:39 crc kubenswrapper[4948]: I0220 08:09:39.729725 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df73228e-df1f-4e9c-aafe-493bb743f98a" path="/var/lib/kubelet/pods/df73228e-df1f-4e9c-aafe-493bb743f98a/volumes" Feb 20 08:09:41 crc kubenswrapper[4948]: I0220 08:09:41.710964 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tvwkh" Feb 20 08:09:41 crc kubenswrapper[4948]: I0220 08:09:41.751385 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" podStartSLOduration=29.751347097 podStartE2EDuration="29.751347097s" podCreationTimestamp="2026-02-20 08:09:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:09:39.482487568 +0000 UTC m=+228.456982428" watchObservedRunningTime="2026-02-20 08:09:41.751347097 +0000 UTC m=+230.725841967" Feb 20 08:09:41 crc kubenswrapper[4948]: I0220 08:09:41.774605 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tvwkh"] Feb 20 08:09:42 crc kubenswrapper[4948]: I0220 08:09:42.441583 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tvwkh" podUID="b6d696cd-f5cf-47e0-af8d-4d20e93bd80d" containerName="registry-server" containerID="cri-o://3eb57d35e2ca5917ed63ba672507ebdc1b840d438dcdc8930f619e051d58336d" gracePeriod=2 Feb 20 08:09:42 crc kubenswrapper[4948]: I0220 08:09:42.877426 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tvwkh" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.045477 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-px4gp\" (UniqueName: \"kubernetes.io/projected/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-kube-api-access-px4gp\") pod \"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d\" (UID: \"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d\") " Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.045608 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-catalog-content\") pod \"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d\" (UID: \"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d\") " Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.045646 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-utilities\") pod \"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d\" (UID: \"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d\") " Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.047600 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-utilities" (OuterVolumeSpecName: "utilities") pod "b6d696cd-f5cf-47e0-af8d-4d20e93bd80d" (UID: "b6d696cd-f5cf-47e0-af8d-4d20e93bd80d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.056118 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-kube-api-access-px4gp" (OuterVolumeSpecName: "kube-api-access-px4gp") pod "b6d696cd-f5cf-47e0-af8d-4d20e93bd80d" (UID: "b6d696cd-f5cf-47e0-af8d-4d20e93bd80d"). InnerVolumeSpecName "kube-api-access-px4gp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.114911 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b6d696cd-f5cf-47e0-af8d-4d20e93bd80d" (UID: "b6d696cd-f5cf-47e0-af8d-4d20e93bd80d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.147152 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-px4gp\" (UniqueName: \"kubernetes.io/projected/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-kube-api-access-px4gp\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.147199 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.147216 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.450513 4948 generic.go:334] "Generic (PLEG): container finished" podID="b6d696cd-f5cf-47e0-af8d-4d20e93bd80d" containerID="3eb57d35e2ca5917ed63ba672507ebdc1b840d438dcdc8930f619e051d58336d" exitCode=0 Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.451188 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tvwkh" event={"ID":"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d","Type":"ContainerDied","Data":"3eb57d35e2ca5917ed63ba672507ebdc1b840d438dcdc8930f619e051d58336d"} Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.451295 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tvwkh" event={"ID":"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d","Type":"ContainerDied","Data":"46adc95f753dc88a78d1ef713a2cbcbf1789c9d4ad7e29e46f1a4a26d4497aeb"} Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.451306 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tvwkh" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.451333 4948 scope.go:117] "RemoveContainer" containerID="3eb57d35e2ca5917ed63ba672507ebdc1b840d438dcdc8930f619e051d58336d" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.488602 4948 scope.go:117] "RemoveContainer" containerID="adec323384ee4b89adc829f6e8cec6c7df59b4806363aa0408583b7df2842893" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.500445 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tvwkh"] Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.512836 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tvwkh"] Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.526251 4948 scope.go:117] "RemoveContainer" containerID="5e2b81703211ae6dafce6a9ca7e54224c7be8c36c969c7e7bf86ed1763e784b1" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.548168 4948 scope.go:117] "RemoveContainer" containerID="3eb57d35e2ca5917ed63ba672507ebdc1b840d438dcdc8930f619e051d58336d" Feb 20 08:09:43 crc kubenswrapper[4948]: E0220 08:09:43.548764 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3eb57d35e2ca5917ed63ba672507ebdc1b840d438dcdc8930f619e051d58336d\": container with ID starting with 3eb57d35e2ca5917ed63ba672507ebdc1b840d438dcdc8930f619e051d58336d not found: ID does not exist" containerID="3eb57d35e2ca5917ed63ba672507ebdc1b840d438dcdc8930f619e051d58336d" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.548827 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3eb57d35e2ca5917ed63ba672507ebdc1b840d438dcdc8930f619e051d58336d"} err="failed to get container status \"3eb57d35e2ca5917ed63ba672507ebdc1b840d438dcdc8930f619e051d58336d\": rpc error: code = NotFound desc = could not find container \"3eb57d35e2ca5917ed63ba672507ebdc1b840d438dcdc8930f619e051d58336d\": container with ID starting with 3eb57d35e2ca5917ed63ba672507ebdc1b840d438dcdc8930f619e051d58336d not found: ID does not exist" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.548866 4948 scope.go:117] "RemoveContainer" containerID="adec323384ee4b89adc829f6e8cec6c7df59b4806363aa0408583b7df2842893" Feb 20 08:09:43 crc kubenswrapper[4948]: E0220 08:09:43.549537 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"adec323384ee4b89adc829f6e8cec6c7df59b4806363aa0408583b7df2842893\": container with ID starting with adec323384ee4b89adc829f6e8cec6c7df59b4806363aa0408583b7df2842893 not found: ID does not exist" containerID="adec323384ee4b89adc829f6e8cec6c7df59b4806363aa0408583b7df2842893" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.549603 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"adec323384ee4b89adc829f6e8cec6c7df59b4806363aa0408583b7df2842893"} err="failed to get container status \"adec323384ee4b89adc829f6e8cec6c7df59b4806363aa0408583b7df2842893\": rpc error: code = NotFound desc = could not find container \"adec323384ee4b89adc829f6e8cec6c7df59b4806363aa0408583b7df2842893\": container with ID starting with adec323384ee4b89adc829f6e8cec6c7df59b4806363aa0408583b7df2842893 not found: ID does not exist" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.549666 4948 scope.go:117] "RemoveContainer" containerID="5e2b81703211ae6dafce6a9ca7e54224c7be8c36c969c7e7bf86ed1763e784b1" Feb 20 08:09:43 crc kubenswrapper[4948]: E0220 08:09:43.550276 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e2b81703211ae6dafce6a9ca7e54224c7be8c36c969c7e7bf86ed1763e784b1\": container with ID starting with 5e2b81703211ae6dafce6a9ca7e54224c7be8c36c969c7e7bf86ed1763e784b1 not found: ID does not exist" containerID="5e2b81703211ae6dafce6a9ca7e54224c7be8c36c969c7e7bf86ed1763e784b1" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.550336 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e2b81703211ae6dafce6a9ca7e54224c7be8c36c969c7e7bf86ed1763e784b1"} err="failed to get container status \"5e2b81703211ae6dafce6a9ca7e54224c7be8c36c969c7e7bf86ed1763e784b1\": rpc error: code = NotFound desc = could not find container \"5e2b81703211ae6dafce6a9ca7e54224c7be8c36c969c7e7bf86ed1763e784b1\": container with ID starting with 5e2b81703211ae6dafce6a9ca7e54224c7be8c36c969c7e7bf86ed1763e784b1 not found: ID does not exist" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.732112 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6d696cd-f5cf-47e0-af8d-4d20e93bd80d" path="/var/lib/kubelet/pods/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d/volumes" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.856146 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-67ptx" Feb 20 08:09:44 crc kubenswrapper[4948]: I0220 08:09:44.297908 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-frznz" Feb 20 08:09:44 crc kubenswrapper[4948]: I0220 08:09:44.342945 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-frznz" Feb 20 08:09:44 crc kubenswrapper[4948]: I0220 08:09:44.515799 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-snc6x" Feb 20 08:09:44 crc kubenswrapper[4948]: I0220 08:09:44.586891 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-snc6x" Feb 20 08:09:46 crc kubenswrapper[4948]: I0220 08:09:46.152795 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-67ptx"] Feb 20 08:09:46 crc kubenswrapper[4948]: I0220 08:09:46.153597 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-67ptx" podUID="b8150552-4914-4e92-b041-56600ed6d2f5" containerName="registry-server" containerID="cri-o://3ece74a70c863ce99b2f2f650f2d87de55dd27277341567aa5ed99e67f8a4dc8" gracePeriod=2 Feb 20 08:09:46 crc kubenswrapper[4948]: I0220 08:09:46.479404 4948 generic.go:334] "Generic (PLEG): container finished" podID="b8150552-4914-4e92-b041-56600ed6d2f5" containerID="3ece74a70c863ce99b2f2f650f2d87de55dd27277341567aa5ed99e67f8a4dc8" exitCode=0 Feb 20 08:09:46 crc kubenswrapper[4948]: I0220 08:09:46.479475 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-67ptx" event={"ID":"b8150552-4914-4e92-b041-56600ed6d2f5","Type":"ContainerDied","Data":"3ece74a70c863ce99b2f2f650f2d87de55dd27277341567aa5ed99e67f8a4dc8"} Feb 20 08:09:46 crc kubenswrapper[4948]: I0220 08:09:46.573639 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-67ptx" Feb 20 08:09:46 crc kubenswrapper[4948]: I0220 08:09:46.697449 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cx26r\" (UniqueName: \"kubernetes.io/projected/b8150552-4914-4e92-b041-56600ed6d2f5-kube-api-access-cx26r\") pod \"b8150552-4914-4e92-b041-56600ed6d2f5\" (UID: \"b8150552-4914-4e92-b041-56600ed6d2f5\") " Feb 20 08:09:46 crc kubenswrapper[4948]: I0220 08:09:46.697530 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8150552-4914-4e92-b041-56600ed6d2f5-utilities\") pod \"b8150552-4914-4e92-b041-56600ed6d2f5\" (UID: \"b8150552-4914-4e92-b041-56600ed6d2f5\") " Feb 20 08:09:46 crc kubenswrapper[4948]: I0220 08:09:46.697616 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8150552-4914-4e92-b041-56600ed6d2f5-catalog-content\") pod \"b8150552-4914-4e92-b041-56600ed6d2f5\" (UID: \"b8150552-4914-4e92-b041-56600ed6d2f5\") " Feb 20 08:09:46 crc kubenswrapper[4948]: I0220 08:09:46.698779 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8150552-4914-4e92-b041-56600ed6d2f5-utilities" (OuterVolumeSpecName: "utilities") pod "b8150552-4914-4e92-b041-56600ed6d2f5" (UID: "b8150552-4914-4e92-b041-56600ed6d2f5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:09:46 crc kubenswrapper[4948]: I0220 08:09:46.704061 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8150552-4914-4e92-b041-56600ed6d2f5-kube-api-access-cx26r" (OuterVolumeSpecName: "kube-api-access-cx26r") pod "b8150552-4914-4e92-b041-56600ed6d2f5" (UID: "b8150552-4914-4e92-b041-56600ed6d2f5"). InnerVolumeSpecName "kube-api-access-cx26r". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:09:46 crc kubenswrapper[4948]: I0220 08:09:46.734109 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8150552-4914-4e92-b041-56600ed6d2f5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b8150552-4914-4e92-b041-56600ed6d2f5" (UID: "b8150552-4914-4e92-b041-56600ed6d2f5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:09:46 crc kubenswrapper[4948]: I0220 08:09:46.751008 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-snc6x"] Feb 20 08:09:46 crc kubenswrapper[4948]: I0220 08:09:46.751218 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-snc6x" podUID="b5ec9802-aa9c-4518-b99a-1673ab75ec50" containerName="registry-server" containerID="cri-o://59bc680a37628dd6760eae32b4fd77ee425353925297c0763e3ca5bf6107c49e" gracePeriod=2 Feb 20 08:09:46 crc kubenswrapper[4948]: I0220 08:09:46.799389 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8150552-4914-4e92-b041-56600ed6d2f5-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:46 crc kubenswrapper[4948]: I0220 08:09:46.799690 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cx26r\" (UniqueName: \"kubernetes.io/projected/b8150552-4914-4e92-b041-56600ed6d2f5-kube-api-access-cx26r\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:46 crc kubenswrapper[4948]: I0220 08:09:46.799709 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8150552-4914-4e92-b041-56600ed6d2f5-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.076729 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-snc6x" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.208281 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5ec9802-aa9c-4518-b99a-1673ab75ec50-utilities\") pod \"b5ec9802-aa9c-4518-b99a-1673ab75ec50\" (UID: \"b5ec9802-aa9c-4518-b99a-1673ab75ec50\") " Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.208733 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5ec9802-aa9c-4518-b99a-1673ab75ec50-catalog-content\") pod \"b5ec9802-aa9c-4518-b99a-1673ab75ec50\" (UID: \"b5ec9802-aa9c-4518-b99a-1673ab75ec50\") " Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.208910 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gc4n5\" (UniqueName: \"kubernetes.io/projected/b5ec9802-aa9c-4518-b99a-1673ab75ec50-kube-api-access-gc4n5\") pod \"b5ec9802-aa9c-4518-b99a-1673ab75ec50\" (UID: \"b5ec9802-aa9c-4518-b99a-1673ab75ec50\") " Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.209297 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5ec9802-aa9c-4518-b99a-1673ab75ec50-utilities" (OuterVolumeSpecName: "utilities") pod "b5ec9802-aa9c-4518-b99a-1673ab75ec50" (UID: "b5ec9802-aa9c-4518-b99a-1673ab75ec50"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.214282 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5ec9802-aa9c-4518-b99a-1673ab75ec50-kube-api-access-gc4n5" (OuterVolumeSpecName: "kube-api-access-gc4n5") pod "b5ec9802-aa9c-4518-b99a-1673ab75ec50" (UID: "b5ec9802-aa9c-4518-b99a-1673ab75ec50"). InnerVolumeSpecName "kube-api-access-gc4n5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.309852 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gc4n5\" (UniqueName: \"kubernetes.io/projected/b5ec9802-aa9c-4518-b99a-1673ab75ec50-kube-api-access-gc4n5\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.309890 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5ec9802-aa9c-4518-b99a-1673ab75ec50-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.340858 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5ec9802-aa9c-4518-b99a-1673ab75ec50-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b5ec9802-aa9c-4518-b99a-1673ab75ec50" (UID: "b5ec9802-aa9c-4518-b99a-1673ab75ec50"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.411644 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5ec9802-aa9c-4518-b99a-1673ab75ec50-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.487085 4948 generic.go:334] "Generic (PLEG): container finished" podID="b5ec9802-aa9c-4518-b99a-1673ab75ec50" containerID="59bc680a37628dd6760eae32b4fd77ee425353925297c0763e3ca5bf6107c49e" exitCode=0 Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.487164 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-snc6x" event={"ID":"b5ec9802-aa9c-4518-b99a-1673ab75ec50","Type":"ContainerDied","Data":"59bc680a37628dd6760eae32b4fd77ee425353925297c0763e3ca5bf6107c49e"} Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.487199 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-snc6x" event={"ID":"b5ec9802-aa9c-4518-b99a-1673ab75ec50","Type":"ContainerDied","Data":"40f046d646c178228156ce2281e80f2e2be9092681446e798f497080898603ae"} Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.487220 4948 scope.go:117] "RemoveContainer" containerID="59bc680a37628dd6760eae32b4fd77ee425353925297c0763e3ca5bf6107c49e" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.487247 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-snc6x" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.489953 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-67ptx" event={"ID":"b8150552-4914-4e92-b041-56600ed6d2f5","Type":"ContainerDied","Data":"b740ba77241ab7ac59567dc96b03935ff231d4a2d8855de4a6ed59ddebaef76b"} Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.490202 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-67ptx" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.508923 4948 scope.go:117] "RemoveContainer" containerID="e970f5a271f9d8c2873e14eacbf1c841136223bb57e91d7713eb158290760f1e" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.529465 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-snc6x"] Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.533309 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-snc6x"] Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.536380 4948 scope.go:117] "RemoveContainer" containerID="5e7272299ae3bd4fd920b69a3700514cf14947047239ae7f3143eafd6f2d9320" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.540906 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-67ptx"] Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.553682 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-67ptx"] Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.564881 4948 scope.go:117] "RemoveContainer" containerID="59bc680a37628dd6760eae32b4fd77ee425353925297c0763e3ca5bf6107c49e" Feb 20 08:09:47 crc kubenswrapper[4948]: E0220 08:09:47.565461 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59bc680a37628dd6760eae32b4fd77ee425353925297c0763e3ca5bf6107c49e\": container with ID starting with 59bc680a37628dd6760eae32b4fd77ee425353925297c0763e3ca5bf6107c49e not found: ID does not exist" containerID="59bc680a37628dd6760eae32b4fd77ee425353925297c0763e3ca5bf6107c49e" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.565502 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59bc680a37628dd6760eae32b4fd77ee425353925297c0763e3ca5bf6107c49e"} err="failed to get container status \"59bc680a37628dd6760eae32b4fd77ee425353925297c0763e3ca5bf6107c49e\": rpc error: code = NotFound desc = could not find container \"59bc680a37628dd6760eae32b4fd77ee425353925297c0763e3ca5bf6107c49e\": container with ID starting with 59bc680a37628dd6760eae32b4fd77ee425353925297c0763e3ca5bf6107c49e not found: ID does not exist" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.565529 4948 scope.go:117] "RemoveContainer" containerID="e970f5a271f9d8c2873e14eacbf1c841136223bb57e91d7713eb158290760f1e" Feb 20 08:09:47 crc kubenswrapper[4948]: E0220 08:09:47.565866 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e970f5a271f9d8c2873e14eacbf1c841136223bb57e91d7713eb158290760f1e\": container with ID starting with e970f5a271f9d8c2873e14eacbf1c841136223bb57e91d7713eb158290760f1e not found: ID does not exist" containerID="e970f5a271f9d8c2873e14eacbf1c841136223bb57e91d7713eb158290760f1e" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.565894 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e970f5a271f9d8c2873e14eacbf1c841136223bb57e91d7713eb158290760f1e"} err="failed to get container status \"e970f5a271f9d8c2873e14eacbf1c841136223bb57e91d7713eb158290760f1e\": rpc error: code = NotFound desc = could not find container \"e970f5a271f9d8c2873e14eacbf1c841136223bb57e91d7713eb158290760f1e\": container with ID starting with e970f5a271f9d8c2873e14eacbf1c841136223bb57e91d7713eb158290760f1e not found: ID does not exist" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.565911 4948 scope.go:117] "RemoveContainer" containerID="5e7272299ae3bd4fd920b69a3700514cf14947047239ae7f3143eafd6f2d9320" Feb 20 08:09:47 crc kubenswrapper[4948]: E0220 08:09:47.566257 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e7272299ae3bd4fd920b69a3700514cf14947047239ae7f3143eafd6f2d9320\": container with ID starting with 5e7272299ae3bd4fd920b69a3700514cf14947047239ae7f3143eafd6f2d9320 not found: ID does not exist" containerID="5e7272299ae3bd4fd920b69a3700514cf14947047239ae7f3143eafd6f2d9320" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.566306 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e7272299ae3bd4fd920b69a3700514cf14947047239ae7f3143eafd6f2d9320"} err="failed to get container status \"5e7272299ae3bd4fd920b69a3700514cf14947047239ae7f3143eafd6f2d9320\": rpc error: code = NotFound desc = could not find container \"5e7272299ae3bd4fd920b69a3700514cf14947047239ae7f3143eafd6f2d9320\": container with ID starting with 5e7272299ae3bd4fd920b69a3700514cf14947047239ae7f3143eafd6f2d9320 not found: ID does not exist" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.566335 4948 scope.go:117] "RemoveContainer" containerID="3ece74a70c863ce99b2f2f650f2d87de55dd27277341567aa5ed99e67f8a4dc8" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.582382 4948 scope.go:117] "RemoveContainer" containerID="e9c973e2d8ac9c5ce88b599645da506dbb8f6dbaabbec1d3e5fca44ff09a3d5c" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.606655 4948 scope.go:117] "RemoveContainer" containerID="24232a1b4e66920a8c674176ee38da24ffd8d17de7ae8a51e030daddaaeb552d" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.734328 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5ec9802-aa9c-4518-b99a-1673ab75ec50" path="/var/lib/kubelet/pods/b5ec9802-aa9c-4518-b99a-1673ab75ec50/volumes" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.735670 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8150552-4914-4e92-b041-56600ed6d2f5" path="/var/lib/kubelet/pods/b8150552-4914-4e92-b041-56600ed6d2f5/volumes" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.967758 4948 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 20 08:09:54 crc kubenswrapper[4948]: E0220 08:09:54.968592 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5ec9802-aa9c-4518-b99a-1673ab75ec50" containerName="extract-utilities" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.968607 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5ec9802-aa9c-4518-b99a-1673ab75ec50" containerName="extract-utilities" Feb 20 08:09:54 crc kubenswrapper[4948]: E0220 08:09:54.968618 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6d696cd-f5cf-47e0-af8d-4d20e93bd80d" containerName="extract-content" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.968625 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6d696cd-f5cf-47e0-af8d-4d20e93bd80d" containerName="extract-content" Feb 20 08:09:54 crc kubenswrapper[4948]: E0220 08:09:54.968638 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8150552-4914-4e92-b041-56600ed6d2f5" containerName="extract-content" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.968648 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8150552-4914-4e92-b041-56600ed6d2f5" containerName="extract-content" Feb 20 08:09:54 crc kubenswrapper[4948]: E0220 08:09:54.968661 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6d696cd-f5cf-47e0-af8d-4d20e93bd80d" containerName="extract-utilities" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.968670 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6d696cd-f5cf-47e0-af8d-4d20e93bd80d" containerName="extract-utilities" Feb 20 08:09:54 crc kubenswrapper[4948]: E0220 08:09:54.968685 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5ec9802-aa9c-4518-b99a-1673ab75ec50" containerName="extract-content" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.968693 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5ec9802-aa9c-4518-b99a-1673ab75ec50" containerName="extract-content" Feb 20 08:09:54 crc kubenswrapper[4948]: E0220 08:09:54.968704 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8150552-4914-4e92-b041-56600ed6d2f5" containerName="extract-utilities" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.968711 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8150552-4914-4e92-b041-56600ed6d2f5" containerName="extract-utilities" Feb 20 08:09:54 crc kubenswrapper[4948]: E0220 08:09:54.968721 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6d696cd-f5cf-47e0-af8d-4d20e93bd80d" containerName="registry-server" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.968728 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6d696cd-f5cf-47e0-af8d-4d20e93bd80d" containerName="registry-server" Feb 20 08:09:54 crc kubenswrapper[4948]: E0220 08:09:54.968739 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5ec9802-aa9c-4518-b99a-1673ab75ec50" containerName="registry-server" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.968747 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5ec9802-aa9c-4518-b99a-1673ab75ec50" containerName="registry-server" Feb 20 08:09:54 crc kubenswrapper[4948]: E0220 08:09:54.968758 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8150552-4914-4e92-b041-56600ed6d2f5" containerName="registry-server" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.968765 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8150552-4914-4e92-b041-56600ed6d2f5" containerName="registry-server" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.968886 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6d696cd-f5cf-47e0-af8d-4d20e93bd80d" containerName="registry-server" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.968904 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8150552-4914-4e92-b041-56600ed6d2f5" containerName="registry-server" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.968913 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5ec9802-aa9c-4518-b99a-1673ab75ec50" containerName="registry-server" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.969289 4948 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.969560 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85" gracePeriod=15 Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.969723 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.970082 4948 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.970107 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244" gracePeriod=15 Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.970205 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61" gracePeriod=15 Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.970173 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556" gracePeriod=15 Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.978080 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a" gracePeriod=15 Feb 20 08:09:54 crc kubenswrapper[4948]: E0220 08:09:54.978190 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.978221 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 20 08:09:54 crc kubenswrapper[4948]: E0220 08:09:54.978241 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.978253 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 20 08:09:54 crc kubenswrapper[4948]: E0220 08:09:54.978268 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.978277 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 20 08:09:54 crc kubenswrapper[4948]: E0220 08:09:54.978292 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.978300 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 20 08:09:54 crc kubenswrapper[4948]: E0220 08:09:54.978310 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.978319 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 20 08:09:54 crc kubenswrapper[4948]: E0220 08:09:54.978330 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.978343 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 20 08:09:54 crc kubenswrapper[4948]: E0220 08:09:54.978357 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.978375 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.978569 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.978580 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.978597 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.978613 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.978625 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.978813 4948 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.978847 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.007587 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.018698 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.018748 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.018776 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.018818 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.018863 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.018896 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.018938 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.018962 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.126909 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.127032 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.127064 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.127097 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.127096 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.127283 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.127319 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.127326 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.127366 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.127368 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.127285 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.127454 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.127576 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.127759 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.127796 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.127817 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.303549 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: W0220 08:09:55.328259 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-c38fe37596ef5bf3d284b1698f5dcb7bfb184524c0da47f1359c8cf118060e1e WatchSource:0}: Error finding container c38fe37596ef5bf3d284b1698f5dcb7bfb184524c0da47f1359c8cf118060e1e: Status 404 returned error can't find the container with id c38fe37596ef5bf3d284b1698f5dcb7bfb184524c0da47f1359c8cf118060e1e Feb 20 08:09:55 crc kubenswrapper[4948]: E0220 08:09:55.331236 4948 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.113:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.1895e61271a3af54 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-20 08:09:55.330510676 +0000 UTC m=+244.305005496,LastTimestamp:2026-02-20 08:09:55.330510676 +0000 UTC m=+244.305005496,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.537450 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"c38fe37596ef5bf3d284b1698f5dcb7bfb184524c0da47f1359c8cf118060e1e"} Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.540507 4948 generic.go:334] "Generic (PLEG): container finished" podID="a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea" containerID="308dd2b5632b47b59d9d195703329e9e330103427c0967b20ee97ba81766b1b3" exitCode=0 Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.540734 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea","Type":"ContainerDied","Data":"308dd2b5632b47b59d9d195703329e9e330103427c0967b20ee97ba81766b1b3"} Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.541601 4948 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.542209 4948 status_manager.go:851] "Failed to get status for pod" podUID="a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.544335 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.546177 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.547261 4948 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61" exitCode=0 Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.547288 4948 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244" exitCode=0 Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.547300 4948 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556" exitCode=0 Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.547309 4948 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a" exitCode=2 Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.547350 4948 scope.go:117] "RemoveContainer" containerID="65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678" Feb 20 08:09:56 crc kubenswrapper[4948]: I0220 08:09:56.557671 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"955d8cda21a3da3121dfa54a2504d8b0f8470dbb104720ff4e180c5ad7329f47"} Feb 20 08:09:56 crc kubenswrapper[4948]: I0220 08:09:56.559110 4948 status_manager.go:851] "Failed to get status for pod" podUID="a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:09:56 crc kubenswrapper[4948]: I0220 08:09:56.559526 4948 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:09:56 crc kubenswrapper[4948]: I0220 08:09:56.561543 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 20 08:09:56 crc kubenswrapper[4948]: I0220 08:09:56.815020 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 20 08:09:56 crc kubenswrapper[4948]: I0220 08:09:56.815849 4948 status_manager.go:851] "Failed to get status for pod" podUID="a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:09:56 crc kubenswrapper[4948]: I0220 08:09:56.816232 4948 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:09:56 crc kubenswrapper[4948]: I0220 08:09:56.955300 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-var-lock\") pod \"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea\" (UID: \"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea\") " Feb 20 08:09:56 crc kubenswrapper[4948]: I0220 08:09:56.955376 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-kubelet-dir\") pod \"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea\" (UID: \"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea\") " Feb 20 08:09:56 crc kubenswrapper[4948]: I0220 08:09:56.955435 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-kube-api-access\") pod \"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea\" (UID: \"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea\") " Feb 20 08:09:56 crc kubenswrapper[4948]: I0220 08:09:56.955440 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-var-lock" (OuterVolumeSpecName: "var-lock") pod "a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea" (UID: "a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:09:56 crc kubenswrapper[4948]: I0220 08:09:56.955603 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea" (UID: "a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:09:56 crc kubenswrapper[4948]: I0220 08:09:56.955816 4948 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-var-lock\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:56 crc kubenswrapper[4948]: I0220 08:09:56.955836 4948 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:56 crc kubenswrapper[4948]: I0220 08:09:56.965087 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea" (UID: "a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.056551 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.367923 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.369119 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.370296 4948 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.370814 4948 status_manager.go:851] "Failed to get status for pod" podUID="a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.371316 4948 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.562764 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.562840 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.562876 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.562913 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.562931 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.562999 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.564628 4948 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.564682 4948 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.564729 4948 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.570057 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea","Type":"ContainerDied","Data":"501f50c7bcee3c3ff4d799d7aadad96a2760427edf1eb9577fa78b654115cf07"} Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.570115 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="501f50c7bcee3c3ff4d799d7aadad96a2760427edf1eb9577fa78b654115cf07" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.570075 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.575816 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.577205 4948 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85" exitCode=0 Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.577304 4948 scope.go:117] "RemoveContainer" containerID="ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.577360 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.594581 4948 status_manager.go:851] "Failed to get status for pod" podUID="a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.595019 4948 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.595669 4948 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.599196 4948 scope.go:117] "RemoveContainer" containerID="fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.604063 4948 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.604490 4948 status_manager.go:851] "Failed to get status for pod" podUID="a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.604727 4948 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.616641 4948 scope.go:117] "RemoveContainer" containerID="b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.633894 4948 scope.go:117] "RemoveContainer" containerID="3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.647761 4948 scope.go:117] "RemoveContainer" containerID="16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.665329 4948 scope.go:117] "RemoveContainer" containerID="1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.690792 4948 scope.go:117] "RemoveContainer" containerID="ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61" Feb 20 08:09:57 crc kubenswrapper[4948]: E0220 08:09:57.697866 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\": container with ID starting with ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61 not found: ID does not exist" containerID="ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.697912 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61"} err="failed to get container status \"ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\": rpc error: code = NotFound desc = could not find container \"ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\": container with ID starting with ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61 not found: ID does not exist" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.697950 4948 scope.go:117] "RemoveContainer" containerID="fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244" Feb 20 08:09:57 crc kubenswrapper[4948]: E0220 08:09:57.699081 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\": container with ID starting with fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244 not found: ID does not exist" containerID="fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.699111 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244"} err="failed to get container status \"fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\": rpc error: code = NotFound desc = could not find container \"fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\": container with ID starting with fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244 not found: ID does not exist" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.699131 4948 scope.go:117] "RemoveContainer" containerID="b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556" Feb 20 08:09:57 crc kubenswrapper[4948]: E0220 08:09:57.699490 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\": container with ID starting with b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556 not found: ID does not exist" containerID="b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.699514 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556"} err="failed to get container status \"b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\": rpc error: code = NotFound desc = could not find container \"b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\": container with ID starting with b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556 not found: ID does not exist" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.699532 4948 scope.go:117] "RemoveContainer" containerID="3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a" Feb 20 08:09:57 crc kubenswrapper[4948]: E0220 08:09:57.699890 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\": container with ID starting with 3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a not found: ID does not exist" containerID="3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.699914 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a"} err="failed to get container status \"3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\": rpc error: code = NotFound desc = could not find container \"3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\": container with ID starting with 3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a not found: ID does not exist" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.699934 4948 scope.go:117] "RemoveContainer" containerID="16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85" Feb 20 08:09:57 crc kubenswrapper[4948]: E0220 08:09:57.700389 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\": container with ID starting with 16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85 not found: ID does not exist" containerID="16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.700415 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85"} err="failed to get container status \"16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\": rpc error: code = NotFound desc = could not find container \"16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\": container with ID starting with 16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85 not found: ID does not exist" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.700432 4948 scope.go:117] "RemoveContainer" containerID="1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b" Feb 20 08:09:57 crc kubenswrapper[4948]: E0220 08:09:57.700875 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\": container with ID starting with 1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b not found: ID does not exist" containerID="1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.700901 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b"} err="failed to get container status \"1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\": rpc error: code = NotFound desc = could not find container \"1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\": container with ID starting with 1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b not found: ID does not exist" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.732045 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Feb 20 08:09:59 crc kubenswrapper[4948]: E0220 08:09:59.535605 4948 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.113:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.1895e61271a3af54 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-20 08:09:55.330510676 +0000 UTC m=+244.305005496,LastTimestamp:2026-02-20 08:09:55.330510676 +0000 UTC m=+244.305005496,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 20 08:10:01 crc kubenswrapper[4948]: I0220 08:10:01.725272 4948 status_manager.go:851] "Failed to get status for pod" podUID="a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:10:01 crc kubenswrapper[4948]: I0220 08:10:01.725522 4948 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:10:03 crc kubenswrapper[4948]: E0220 08:10:03.246370 4948 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:10:03 crc kubenswrapper[4948]: E0220 08:10:03.247603 4948 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:10:03 crc kubenswrapper[4948]: E0220 08:10:03.248481 4948 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:10:03 crc kubenswrapper[4948]: E0220 08:10:03.248964 4948 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:10:03 crc kubenswrapper[4948]: E0220 08:10:03.249516 4948 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:10:03 crc kubenswrapper[4948]: I0220 08:10:03.249572 4948 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Feb 20 08:10:03 crc kubenswrapper[4948]: E0220 08:10:03.250044 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" interval="200ms" Feb 20 08:10:03 crc kubenswrapper[4948]: E0220 08:10:03.450877 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" interval="400ms" Feb 20 08:10:03 crc kubenswrapper[4948]: E0220 08:10:03.852566 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" interval="800ms" Feb 20 08:10:04 crc kubenswrapper[4948]: E0220 08:10:04.653402 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" interval="1.6s" Feb 20 08:10:06 crc kubenswrapper[4948]: E0220 08:10:06.254096 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" interval="3.2s" Feb 20 08:10:06 crc kubenswrapper[4948]: I0220 08:10:06.722295 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:10:06 crc kubenswrapper[4948]: I0220 08:10:06.723388 4948 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:10:06 crc kubenswrapper[4948]: I0220 08:10:06.723700 4948 status_manager.go:851] "Failed to get status for pod" podUID="a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:10:06 crc kubenswrapper[4948]: I0220 08:10:06.751171 4948 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4bc8c9ea-350e-45e5-af21-f268fb6cbffd" Feb 20 08:10:06 crc kubenswrapper[4948]: I0220 08:10:06.751219 4948 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4bc8c9ea-350e-45e5-af21-f268fb6cbffd" Feb 20 08:10:06 crc kubenswrapper[4948]: E0220 08:10:06.751577 4948 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:10:06 crc kubenswrapper[4948]: I0220 08:10:06.752197 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:10:07 crc kubenswrapper[4948]: I0220 08:10:07.638745 4948 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="cf3409460deb96ddc07d331994e6e633d6585d5115e12c4a1dc7cdd400df8864" exitCode=0 Feb 20 08:10:07 crc kubenswrapper[4948]: I0220 08:10:07.639061 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"cf3409460deb96ddc07d331994e6e633d6585d5115e12c4a1dc7cdd400df8864"} Feb 20 08:10:07 crc kubenswrapper[4948]: I0220 08:10:07.639099 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b459704b071ca4989d633d704fc1c963dfa6f5815ee2bbd817deffc4f13de188"} Feb 20 08:10:07 crc kubenswrapper[4948]: I0220 08:10:07.639427 4948 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4bc8c9ea-350e-45e5-af21-f268fb6cbffd" Feb 20 08:10:07 crc kubenswrapper[4948]: I0220 08:10:07.639447 4948 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4bc8c9ea-350e-45e5-af21-f268fb6cbffd" Feb 20 08:10:07 crc kubenswrapper[4948]: E0220 08:10:07.640320 4948 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:10:07 crc kubenswrapper[4948]: I0220 08:10:07.640424 4948 status_manager.go:851] "Failed to get status for pod" podUID="a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:10:07 crc kubenswrapper[4948]: I0220 08:10:07.642255 4948 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:10:08 crc kubenswrapper[4948]: I0220 08:10:08.658504 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d146155213dd15f0975d2776ef634cc79eded90bb9335fff45469a3710315ccf"} Feb 20 08:10:08 crc kubenswrapper[4948]: I0220 08:10:08.659012 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"cfdeaaf2a6e0394e1775d2d9858b907df2a41e8764c775426f1360d3a27b4a88"} Feb 20 08:10:08 crc kubenswrapper[4948]: I0220 08:10:08.659023 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"fd7dd20c32f883e583e8daaaa5e8a02aaf07174aa43d4c207aa9a8b98bfee476"} Feb 20 08:10:09 crc kubenswrapper[4948]: I0220 08:10:09.666753 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"8b38b70f3c4ebf804f8bcc0ad1c39d452ca909f1a7cf2c456546f3d0743eb2ce"} Feb 20 08:10:09 crc kubenswrapper[4948]: I0220 08:10:09.667095 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:10:09 crc kubenswrapper[4948]: I0220 08:10:09.667098 4948 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4bc8c9ea-350e-45e5-af21-f268fb6cbffd" Feb 20 08:10:09 crc kubenswrapper[4948]: I0220 08:10:09.667128 4948 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4bc8c9ea-350e-45e5-af21-f268fb6cbffd" Feb 20 08:10:09 crc kubenswrapper[4948]: I0220 08:10:09.667107 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"fb540300e2f72729de60323490ef76783142a419cc9984ab599377a9b2414525"} Feb 20 08:10:09 crc kubenswrapper[4948]: I0220 08:10:09.670616 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 20 08:10:09 crc kubenswrapper[4948]: I0220 08:10:09.670740 4948 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4" exitCode=1 Feb 20 08:10:09 crc kubenswrapper[4948]: I0220 08:10:09.670815 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4"} Feb 20 08:10:09 crc kubenswrapper[4948]: I0220 08:10:09.671288 4948 scope.go:117] "RemoveContainer" containerID="493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4" Feb 20 08:10:10 crc kubenswrapper[4948]: I0220 08:10:10.677924 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 20 08:10:10 crc kubenswrapper[4948]: I0220 08:10:10.678005 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d2fe425396bbcb915a7989ce92d49c39314fbf78eae93aca7ab82235535ad6f5"} Feb 20 08:10:11 crc kubenswrapper[4948]: I0220 08:10:11.254879 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:10:11 crc kubenswrapper[4948]: I0220 08:10:11.753553 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:10:11 crc kubenswrapper[4948]: I0220 08:10:11.753907 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:10:11 crc kubenswrapper[4948]: I0220 08:10:11.762904 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:10:14 crc kubenswrapper[4948]: I0220 08:10:14.678729 4948 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:10:14 crc kubenswrapper[4948]: I0220 08:10:14.706110 4948 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4bc8c9ea-350e-45e5-af21-f268fb6cbffd" Feb 20 08:10:14 crc kubenswrapper[4948]: I0220 08:10:14.706148 4948 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4bc8c9ea-350e-45e5-af21-f268fb6cbffd" Feb 20 08:10:14 crc kubenswrapper[4948]: I0220 08:10:14.714398 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:10:14 crc kubenswrapper[4948]: I0220 08:10:14.773845 4948 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="dd37c39b-433d-4d22-aa05-3bf01b5813fe" Feb 20 08:10:14 crc kubenswrapper[4948]: I0220 08:10:14.970303 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:10:14 crc kubenswrapper[4948]: I0220 08:10:14.973783 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:10:15 crc kubenswrapper[4948]: I0220 08:10:15.713380 4948 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4bc8c9ea-350e-45e5-af21-f268fb6cbffd" Feb 20 08:10:15 crc kubenswrapper[4948]: I0220 08:10:15.713420 4948 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4bc8c9ea-350e-45e5-af21-f268fb6cbffd" Feb 20 08:10:15 crc kubenswrapper[4948]: I0220 08:10:15.718861 4948 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="dd37c39b-433d-4d22-aa05-3bf01b5813fe" Feb 20 08:10:21 crc kubenswrapper[4948]: I0220 08:10:21.262879 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:10:24 crc kubenswrapper[4948]: I0220 08:10:24.437752 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 20 08:10:25 crc kubenswrapper[4948]: I0220 08:10:25.424337 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 20 08:10:25 crc kubenswrapper[4948]: I0220 08:10:25.796461 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 20 08:10:25 crc kubenswrapper[4948]: I0220 08:10:25.980069 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 20 08:10:26 crc kubenswrapper[4948]: I0220 08:10:26.215472 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 20 08:10:26 crc kubenswrapper[4948]: I0220 08:10:26.262062 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 20 08:10:26 crc kubenswrapper[4948]: I0220 08:10:26.344362 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 20 08:10:26 crc kubenswrapper[4948]: I0220 08:10:26.499819 4948 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 20 08:10:26 crc kubenswrapper[4948]: I0220 08:10:26.802384 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 20 08:10:26 crc kubenswrapper[4948]: I0220 08:10:26.846803 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 20 08:10:27 crc kubenswrapper[4948]: I0220 08:10:27.024161 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 20 08:10:27 crc kubenswrapper[4948]: I0220 08:10:27.024503 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 20 08:10:27 crc kubenswrapper[4948]: I0220 08:10:27.283966 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 20 08:10:27 crc kubenswrapper[4948]: I0220 08:10:27.710563 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 20 08:10:27 crc kubenswrapper[4948]: I0220 08:10:27.745821 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 20 08:10:27 crc kubenswrapper[4948]: I0220 08:10:27.769348 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 20 08:10:27 crc kubenswrapper[4948]: I0220 08:10:27.969627 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 20 08:10:28 crc kubenswrapper[4948]: I0220 08:10:28.152916 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 20 08:10:28 crc kubenswrapper[4948]: I0220 08:10:28.160730 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 20 08:10:28 crc kubenswrapper[4948]: I0220 08:10:28.225312 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 20 08:10:28 crc kubenswrapper[4948]: I0220 08:10:28.255887 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 20 08:10:28 crc kubenswrapper[4948]: I0220 08:10:28.364838 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 20 08:10:28 crc kubenswrapper[4948]: I0220 08:10:28.490718 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 20 08:10:28 crc kubenswrapper[4948]: I0220 08:10:28.615037 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 20 08:10:28 crc kubenswrapper[4948]: I0220 08:10:28.627027 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 20 08:10:28 crc kubenswrapper[4948]: I0220 08:10:28.701322 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 20 08:10:28 crc kubenswrapper[4948]: I0220 08:10:28.829250 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 20 08:10:28 crc kubenswrapper[4948]: I0220 08:10:28.885207 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 20 08:10:28 crc kubenswrapper[4948]: I0220 08:10:28.988261 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 20 08:10:29 crc kubenswrapper[4948]: I0220 08:10:29.020194 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 20 08:10:29 crc kubenswrapper[4948]: I0220 08:10:29.030615 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 20 08:10:29 crc kubenswrapper[4948]: I0220 08:10:29.140142 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 20 08:10:29 crc kubenswrapper[4948]: I0220 08:10:29.187003 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 20 08:10:29 crc kubenswrapper[4948]: I0220 08:10:29.190649 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 20 08:10:29 crc kubenswrapper[4948]: I0220 08:10:29.269838 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 20 08:10:29 crc kubenswrapper[4948]: I0220 08:10:29.384116 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 20 08:10:29 crc kubenswrapper[4948]: I0220 08:10:29.388528 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 20 08:10:29 crc kubenswrapper[4948]: I0220 08:10:29.418782 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 20 08:10:29 crc kubenswrapper[4948]: I0220 08:10:29.523323 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 20 08:10:29 crc kubenswrapper[4948]: I0220 08:10:29.632662 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 20 08:10:29 crc kubenswrapper[4948]: I0220 08:10:29.757396 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 20 08:10:29 crc kubenswrapper[4948]: I0220 08:10:29.812752 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 20 08:10:29 crc kubenswrapper[4948]: I0220 08:10:29.958845 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 20 08:10:29 crc kubenswrapper[4948]: I0220 08:10:29.982532 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 20 08:10:30 crc kubenswrapper[4948]: I0220 08:10:30.037122 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 20 08:10:30 crc kubenswrapper[4948]: I0220 08:10:30.129737 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 20 08:10:30 crc kubenswrapper[4948]: I0220 08:10:30.159252 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 20 08:10:30 crc kubenswrapper[4948]: I0220 08:10:30.286953 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 20 08:10:30 crc kubenswrapper[4948]: I0220 08:10:30.323060 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 20 08:10:30 crc kubenswrapper[4948]: I0220 08:10:30.339277 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 20 08:10:30 crc kubenswrapper[4948]: I0220 08:10:30.364709 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 20 08:10:30 crc kubenswrapper[4948]: I0220 08:10:30.453608 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 20 08:10:30 crc kubenswrapper[4948]: I0220 08:10:30.471895 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 20 08:10:30 crc kubenswrapper[4948]: I0220 08:10:30.683431 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 20 08:10:30 crc kubenswrapper[4948]: I0220 08:10:30.707998 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.088697 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.099003 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.142022 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.175650 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.240430 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.324698 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.359883 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.458098 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.533131 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.541370 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.560504 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.587631 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.622136 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.630156 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.726950 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.732406 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.800728 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.929772 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.994756 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.999901 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.116050 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.187827 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.197600 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.199565 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.216050 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.290106 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.294618 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.341361 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.352656 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.429073 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.515230 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.636898 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.694724 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.717833 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.775956 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.918805 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.962093 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.963837 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.973752 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 20 08:10:33 crc kubenswrapper[4948]: I0220 08:10:33.167897 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 20 08:10:33 crc kubenswrapper[4948]: I0220 08:10:33.170726 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 20 08:10:33 crc kubenswrapper[4948]: I0220 08:10:33.240062 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 20 08:10:33 crc kubenswrapper[4948]: I0220 08:10:33.299781 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 20 08:10:33 crc kubenswrapper[4948]: I0220 08:10:33.310708 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 20 08:10:33 crc kubenswrapper[4948]: I0220 08:10:33.363302 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 20 08:10:33 crc kubenswrapper[4948]: I0220 08:10:33.368475 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 20 08:10:33 crc kubenswrapper[4948]: I0220 08:10:33.403932 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 20 08:10:33 crc kubenswrapper[4948]: I0220 08:10:33.545097 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 20 08:10:33 crc kubenswrapper[4948]: I0220 08:10:33.562661 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 20 08:10:33 crc kubenswrapper[4948]: I0220 08:10:33.679161 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 20 08:10:33 crc kubenswrapper[4948]: I0220 08:10:33.716370 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 20 08:10:33 crc kubenswrapper[4948]: I0220 08:10:33.852172 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 20 08:10:33 crc kubenswrapper[4948]: I0220 08:10:33.891157 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 20 08:10:33 crc kubenswrapper[4948]: I0220 08:10:33.900049 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.019988 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.060469 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.122495 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.200219 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.208492 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.221587 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.307244 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.310843 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.330932 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.357798 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.364653 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.409812 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.455752 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.474355 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.566512 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.673849 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.705399 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.706100 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.849066 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.853987 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.854934 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.894550 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.897774 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.911057 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.975683 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.029162 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.127953 4948 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.131628 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.188138 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.205617 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.213716 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.243451 4948 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.249248 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=41.24921699 podStartE2EDuration="41.24921699s" podCreationTimestamp="2026-02-20 08:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:10:14.729957127 +0000 UTC m=+263.704451987" watchObservedRunningTime="2026-02-20 08:10:35.24921699 +0000 UTC m=+284.223711810" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.251614 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.253302 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.253375 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.277036 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=21.277005192 podStartE2EDuration="21.277005192s" podCreationTimestamp="2026-02-20 08:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:10:35.276289474 +0000 UTC m=+284.250784304" watchObservedRunningTime="2026-02-20 08:10:35.277005192 +0000 UTC m=+284.251500042" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.306696 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.350956 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.376296 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.392327 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.400903 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.459217 4948 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.462266 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.525945 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.526515 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.591485 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.592350 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.594073 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.677857 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.748035 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.773585 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.840124 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.076162 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.117214 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.156611 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.277186 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.277298 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.329247 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.338800 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.369185 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.396323 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.408452 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.491020 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.555592 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.564956 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.659437 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.696127 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.717185 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.759805 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.775891 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.812825 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.882830 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.883804 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.002211 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.013915 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.040646 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.055578 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.097206 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.107390 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.165690 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.170251 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.172215 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.306023 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.342620 4948 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.342923 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://955d8cda21a3da3121dfa54a2504d8b0f8470dbb104720ff4e180c5ad7329f47" gracePeriod=5 Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.348251 4948 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.456201 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.481828 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.693955 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.788890 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.849490 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 20 08:10:38 crc kubenswrapper[4948]: I0220 08:10:38.100639 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 20 08:10:38 crc kubenswrapper[4948]: I0220 08:10:38.154676 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 20 08:10:38 crc kubenswrapper[4948]: I0220 08:10:38.191527 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 20 08:10:38 crc kubenswrapper[4948]: I0220 08:10:38.202256 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 20 08:10:38 crc kubenswrapper[4948]: I0220 08:10:38.203174 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 20 08:10:38 crc kubenswrapper[4948]: I0220 08:10:38.286744 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 20 08:10:38 crc kubenswrapper[4948]: I0220 08:10:38.379135 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 20 08:10:38 crc kubenswrapper[4948]: I0220 08:10:38.453529 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 20 08:10:38 crc kubenswrapper[4948]: I0220 08:10:38.469400 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 20 08:10:38 crc kubenswrapper[4948]: I0220 08:10:38.549234 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 20 08:10:38 crc kubenswrapper[4948]: I0220 08:10:38.617320 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 20 08:10:38 crc kubenswrapper[4948]: I0220 08:10:38.620229 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 20 08:10:38 crc kubenswrapper[4948]: I0220 08:10:38.672865 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 20 08:10:38 crc kubenswrapper[4948]: I0220 08:10:38.845388 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 20 08:10:38 crc kubenswrapper[4948]: I0220 08:10:38.871067 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 20 08:10:38 crc kubenswrapper[4948]: I0220 08:10:38.910714 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 20 08:10:38 crc kubenswrapper[4948]: I0220 08:10:38.950128 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.026824 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nrhl6"] Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.027084 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nrhl6" podUID="f75a50cf-7dae-420b-a00d-671ede3cb6f7" containerName="registry-server" containerID="cri-o://51c14bb6285644fd84e90599149f7f330a4395b58ff4509817fbdce179d6470b" gracePeriod=30 Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.046320 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n8d7m"] Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.046655 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-n8d7m" podUID="00f4b4c1-af36-4ccd-a3d1-dd6df1186338" containerName="registry-server" containerID="cri-o://acfb3ae1070bc662949f97e63b0d95f3a1957ddfef18898cde66fcbe8186d502" gracePeriod=30 Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.054164 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sjv8t"] Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.054679 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" podUID="95a270e5-7a08-4238-a65d-96986f2d2c39" containerName="marketplace-operator" containerID="cri-o://15687151f4ed59c06cf2071d3204504981cf4a0c6cdb40d25a5dbd083313ed65" gracePeriod=30 Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.066041 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c9z42"] Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.066269 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-c9z42" podUID="c53ddd33-0f9f-4794-b346-7d48a6c09c9b" containerName="registry-server" containerID="cri-o://1c4c2fe97c46c10efdb2a7e794505136dc520212167c78787ccbac94170a83c0" gracePeriod=30 Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.072277 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-frznz"] Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.073426 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-frznz" podUID="dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" containerName="registry-server" containerID="cri-o://fb097a14ed0ab7d5af6abab587a1c01e68baf8b43ccad645f65796167b0c90bd" gracePeriod=30 Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.110275 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tbqzs"] Feb 20 08:10:39 crc kubenswrapper[4948]: E0220 08:10:39.110573 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea" containerName="installer" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.110602 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea" containerName="installer" Feb 20 08:10:39 crc kubenswrapper[4948]: E0220 08:10:39.110663 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.110677 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.110816 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.110849 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea" containerName="installer" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.111467 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tbqzs" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.116316 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.123175 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tbqzs"] Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.185083 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.222385 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.225387 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1e3d2cd4-4311-4454-b257-20a3caf243b3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tbqzs\" (UID: \"1e3d2cd4-4311-4454-b257-20a3caf243b3\") " pod="openshift-marketplace/marketplace-operator-79b997595-tbqzs" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.225440 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1e3d2cd4-4311-4454-b257-20a3caf243b3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tbqzs\" (UID: \"1e3d2cd4-4311-4454-b257-20a3caf243b3\") " pod="openshift-marketplace/marketplace-operator-79b997595-tbqzs" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.225460 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfqrz\" (UniqueName: \"kubernetes.io/projected/1e3d2cd4-4311-4454-b257-20a3caf243b3-kube-api-access-cfqrz\") pod \"marketplace-operator-79b997595-tbqzs\" (UID: \"1e3d2cd4-4311-4454-b257-20a3caf243b3\") " pod="openshift-marketplace/marketplace-operator-79b997595-tbqzs" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.327083 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1e3d2cd4-4311-4454-b257-20a3caf243b3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tbqzs\" (UID: \"1e3d2cd4-4311-4454-b257-20a3caf243b3\") " pod="openshift-marketplace/marketplace-operator-79b997595-tbqzs" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.327129 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1e3d2cd4-4311-4454-b257-20a3caf243b3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tbqzs\" (UID: \"1e3d2cd4-4311-4454-b257-20a3caf243b3\") " pod="openshift-marketplace/marketplace-operator-79b997595-tbqzs" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.327154 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfqrz\" (UniqueName: \"kubernetes.io/projected/1e3d2cd4-4311-4454-b257-20a3caf243b3-kube-api-access-cfqrz\") pod \"marketplace-operator-79b997595-tbqzs\" (UID: \"1e3d2cd4-4311-4454-b257-20a3caf243b3\") " pod="openshift-marketplace/marketplace-operator-79b997595-tbqzs" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.328432 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1e3d2cd4-4311-4454-b257-20a3caf243b3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tbqzs\" (UID: \"1e3d2cd4-4311-4454-b257-20a3caf243b3\") " pod="openshift-marketplace/marketplace-operator-79b997595-tbqzs" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.334798 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1e3d2cd4-4311-4454-b257-20a3caf243b3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tbqzs\" (UID: \"1e3d2cd4-4311-4454-b257-20a3caf243b3\") " pod="openshift-marketplace/marketplace-operator-79b997595-tbqzs" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.340629 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.343516 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfqrz\" (UniqueName: \"kubernetes.io/projected/1e3d2cd4-4311-4454-b257-20a3caf243b3-kube-api-access-cfqrz\") pod \"marketplace-operator-79b997595-tbqzs\" (UID: \"1e3d2cd4-4311-4454-b257-20a3caf243b3\") " pod="openshift-marketplace/marketplace-operator-79b997595-tbqzs" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.354171 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.373929 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.452836 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tbqzs" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.464328 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.502785 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-frznz" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.508812 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c9z42" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.512671 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n8d7m" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.517569 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nrhl6" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.581243 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.593418 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.616421 4948 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.626022 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.632123 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-utilities\") pod \"c53ddd33-0f9f-4794-b346-7d48a6c09c9b\" (UID: \"c53ddd33-0f9f-4794-b346-7d48a6c09c9b\") " Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.632178 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55q6w\" (UniqueName: \"kubernetes.io/projected/95a270e5-7a08-4238-a65d-96986f2d2c39-kube-api-access-55q6w\") pod \"95a270e5-7a08-4238-a65d-96986f2d2c39\" (UID: \"95a270e5-7a08-4238-a65d-96986f2d2c39\") " Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.632211 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxq9k\" (UniqueName: \"kubernetes.io/projected/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-kube-api-access-kxq9k\") pod \"00f4b4c1-af36-4ccd-a3d1-dd6df1186338\" (UID: \"00f4b4c1-af36-4ccd-a3d1-dd6df1186338\") " Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.632231 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-catalog-content\") pod \"00f4b4c1-af36-4ccd-a3d1-dd6df1186338\" (UID: \"00f4b4c1-af36-4ccd-a3d1-dd6df1186338\") " Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.632250 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqtvv\" (UniqueName: \"kubernetes.io/projected/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-kube-api-access-cqtvv\") pod \"c53ddd33-0f9f-4794-b346-7d48a6c09c9b\" (UID: \"c53ddd33-0f9f-4794-b346-7d48a6c09c9b\") " Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.632267 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-catalog-content\") pod \"c53ddd33-0f9f-4794-b346-7d48a6c09c9b\" (UID: \"c53ddd33-0f9f-4794-b346-7d48a6c09c9b\") " Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.632286 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-catalog-content\") pod \"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9\" (UID: \"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9\") " Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.632322 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-utilities\") pod \"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9\" (UID: \"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9\") " Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.632342 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f75a50cf-7dae-420b-a00d-671ede3cb6f7-utilities\") pod \"f75a50cf-7dae-420b-a00d-671ede3cb6f7\" (UID: \"f75a50cf-7dae-420b-a00d-671ede3cb6f7\") " Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.632356 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqs4x\" (UniqueName: \"kubernetes.io/projected/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-kube-api-access-hqs4x\") pod \"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9\" (UID: \"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9\") " Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.632374 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/95a270e5-7a08-4238-a65d-96986f2d2c39-marketplace-operator-metrics\") pod \"95a270e5-7a08-4238-a65d-96986f2d2c39\" (UID: \"95a270e5-7a08-4238-a65d-96986f2d2c39\") " Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.632413 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nc669\" (UniqueName: \"kubernetes.io/projected/f75a50cf-7dae-420b-a00d-671ede3cb6f7-kube-api-access-nc669\") pod \"f75a50cf-7dae-420b-a00d-671ede3cb6f7\" (UID: \"f75a50cf-7dae-420b-a00d-671ede3cb6f7\") " Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.632443 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f75a50cf-7dae-420b-a00d-671ede3cb6f7-catalog-content\") pod \"f75a50cf-7dae-420b-a00d-671ede3cb6f7\" (UID: \"f75a50cf-7dae-420b-a00d-671ede3cb6f7\") " Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.632465 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/95a270e5-7a08-4238-a65d-96986f2d2c39-marketplace-trusted-ca\") pod \"95a270e5-7a08-4238-a65d-96986f2d2c39\" (UID: \"95a270e5-7a08-4238-a65d-96986f2d2c39\") " Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.632481 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-utilities\") pod \"00f4b4c1-af36-4ccd-a3d1-dd6df1186338\" (UID: \"00f4b4c1-af36-4ccd-a3d1-dd6df1186338\") " Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.633587 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-utilities" (OuterVolumeSpecName: "utilities") pod "c53ddd33-0f9f-4794-b346-7d48a6c09c9b" (UID: "c53ddd33-0f9f-4794-b346-7d48a6c09c9b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.634332 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-utilities" (OuterVolumeSpecName: "utilities") pod "dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" (UID: "dc2fe1f3-084a-4e45-b64d-f385fb28d7e9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.634573 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f75a50cf-7dae-420b-a00d-671ede3cb6f7-utilities" (OuterVolumeSpecName: "utilities") pod "f75a50cf-7dae-420b-a00d-671ede3cb6f7" (UID: "f75a50cf-7dae-420b-a00d-671ede3cb6f7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.635357 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95a270e5-7a08-4238-a65d-96986f2d2c39-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "95a270e5-7a08-4238-a65d-96986f2d2c39" (UID: "95a270e5-7a08-4238-a65d-96986f2d2c39"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.635452 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-utilities" (OuterVolumeSpecName: "utilities") pod "00f4b4c1-af36-4ccd-a3d1-dd6df1186338" (UID: "00f4b4c1-af36-4ccd-a3d1-dd6df1186338"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.638495 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f75a50cf-7dae-420b-a00d-671ede3cb6f7-kube-api-access-nc669" (OuterVolumeSpecName: "kube-api-access-nc669") pod "f75a50cf-7dae-420b-a00d-671ede3cb6f7" (UID: "f75a50cf-7dae-420b-a00d-671ede3cb6f7"). InnerVolumeSpecName "kube-api-access-nc669". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.638688 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-kube-api-access-cqtvv" (OuterVolumeSpecName: "kube-api-access-cqtvv") pod "c53ddd33-0f9f-4794-b346-7d48a6c09c9b" (UID: "c53ddd33-0f9f-4794-b346-7d48a6c09c9b"). InnerVolumeSpecName "kube-api-access-cqtvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.638821 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-kube-api-access-hqs4x" (OuterVolumeSpecName: "kube-api-access-hqs4x") pod "dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" (UID: "dc2fe1f3-084a-4e45-b64d-f385fb28d7e9"). InnerVolumeSpecName "kube-api-access-hqs4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.641780 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-kube-api-access-kxq9k" (OuterVolumeSpecName: "kube-api-access-kxq9k") pod "00f4b4c1-af36-4ccd-a3d1-dd6df1186338" (UID: "00f4b4c1-af36-4ccd-a3d1-dd6df1186338"). InnerVolumeSpecName "kube-api-access-kxq9k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.651549 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95a270e5-7a08-4238-a65d-96986f2d2c39-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "95a270e5-7a08-4238-a65d-96986f2d2c39" (UID: "95a270e5-7a08-4238-a65d-96986f2d2c39"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.662348 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95a270e5-7a08-4238-a65d-96986f2d2c39-kube-api-access-55q6w" (OuterVolumeSpecName: "kube-api-access-55q6w") pod "95a270e5-7a08-4238-a65d-96986f2d2c39" (UID: "95a270e5-7a08-4238-a65d-96986f2d2c39"). InnerVolumeSpecName "kube-api-access-55q6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.664879 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c53ddd33-0f9f-4794-b346-7d48a6c09c9b" (UID: "c53ddd33-0f9f-4794-b346-7d48a6c09c9b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.676199 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.689313 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tbqzs"] Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.724894 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "00f4b4c1-af36-4ccd-a3d1-dd6df1186338" (UID: "00f4b4c1-af36-4ccd-a3d1-dd6df1186338"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.733434 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nc669\" (UniqueName: \"kubernetes.io/projected/f75a50cf-7dae-420b-a00d-671ede3cb6f7-kube-api-access-nc669\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.733767 4948 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/95a270e5-7a08-4238-a65d-96986f2d2c39-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.733777 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.733787 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.733799 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55q6w\" (UniqueName: \"kubernetes.io/projected/95a270e5-7a08-4238-a65d-96986f2d2c39-kube-api-access-55q6w\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.733809 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxq9k\" (UniqueName: \"kubernetes.io/projected/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-kube-api-access-kxq9k\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.733818 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.733828 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.733836 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqtvv\" (UniqueName: \"kubernetes.io/projected/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-kube-api-access-cqtvv\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.733844 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.733853 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f75a50cf-7dae-420b-a00d-671ede3cb6f7-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.733861 4948 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/95a270e5-7a08-4238-a65d-96986f2d2c39-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.733891 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqs4x\" (UniqueName: \"kubernetes.io/projected/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-kube-api-access-hqs4x\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.738284 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f75a50cf-7dae-420b-a00d-671ede3cb6f7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f75a50cf-7dae-420b-a00d-671ede3cb6f7" (UID: "f75a50cf-7dae-420b-a00d-671ede3cb6f7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.780533 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" (UID: "dc2fe1f3-084a-4e45-b64d-f385fb28d7e9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.835502 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.835601 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f75a50cf-7dae-420b-a00d-671ede3cb6f7-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.862792 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.875741 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.885233 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.885326 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9z42" event={"ID":"c53ddd33-0f9f-4794-b346-7d48a6c09c9b","Type":"ContainerDied","Data":"1c4c2fe97c46c10efdb2a7e794505136dc520212167c78787ccbac94170a83c0"} Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.885369 4948 scope.go:117] "RemoveContainer" containerID="1c4c2fe97c46c10efdb2a7e794505136dc520212167c78787ccbac94170a83c0" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.885372 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c9z42" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.885299 4948 generic.go:334] "Generic (PLEG): container finished" podID="c53ddd33-0f9f-4794-b346-7d48a6c09c9b" containerID="1c4c2fe97c46c10efdb2a7e794505136dc520212167c78787ccbac94170a83c0" exitCode=0 Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.885477 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9z42" event={"ID":"c53ddd33-0f9f-4794-b346-7d48a6c09c9b","Type":"ContainerDied","Data":"9a7c5f1538f439aba0332cfbd20f5f653d7b240472d961eff61974598f86365b"} Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.890474 4948 generic.go:334] "Generic (PLEG): container finished" podID="dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" containerID="fb097a14ed0ab7d5af6abab587a1c01e68baf8b43ccad645f65796167b0c90bd" exitCode=0 Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.890539 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frznz" event={"ID":"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9","Type":"ContainerDied","Data":"fb097a14ed0ab7d5af6abab587a1c01e68baf8b43ccad645f65796167b0c90bd"} Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.890568 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frznz" event={"ID":"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9","Type":"ContainerDied","Data":"bc370df4b0b552966157b6ac4af296601fb860ed2508da9c4291e9924ec1e9cb"} Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.890626 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-frznz" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.897407 4948 generic.go:334] "Generic (PLEG): container finished" podID="95a270e5-7a08-4238-a65d-96986f2d2c39" containerID="15687151f4ed59c06cf2071d3204504981cf4a0c6cdb40d25a5dbd083313ed65" exitCode=0 Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.897671 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.899224 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" event={"ID":"95a270e5-7a08-4238-a65d-96986f2d2c39","Type":"ContainerDied","Data":"15687151f4ed59c06cf2071d3204504981cf4a0c6cdb40d25a5dbd083313ed65"} Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.899306 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" event={"ID":"95a270e5-7a08-4238-a65d-96986f2d2c39","Type":"ContainerDied","Data":"936301da215b82b117fb711a926cf169f025d16470e0871be5907e8efa5aec8e"} Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.903654 4948 generic.go:334] "Generic (PLEG): container finished" podID="00f4b4c1-af36-4ccd-a3d1-dd6df1186338" containerID="acfb3ae1070bc662949f97e63b0d95f3a1957ddfef18898cde66fcbe8186d502" exitCode=0 Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.903738 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n8d7m" event={"ID":"00f4b4c1-af36-4ccd-a3d1-dd6df1186338","Type":"ContainerDied","Data":"acfb3ae1070bc662949f97e63b0d95f3a1957ddfef18898cde66fcbe8186d502"} Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.903769 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n8d7m" event={"ID":"00f4b4c1-af36-4ccd-a3d1-dd6df1186338","Type":"ContainerDied","Data":"33b053c8152742a922c7d8709831de18b8993625ae7594197966e98f173b1174"} Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.903848 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n8d7m" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.910309 4948 generic.go:334] "Generic (PLEG): container finished" podID="f75a50cf-7dae-420b-a00d-671ede3cb6f7" containerID="51c14bb6285644fd84e90599149f7f330a4395b58ff4509817fbdce179d6470b" exitCode=0 Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.910374 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nrhl6" event={"ID":"f75a50cf-7dae-420b-a00d-671ede3cb6f7","Type":"ContainerDied","Data":"51c14bb6285644fd84e90599149f7f330a4395b58ff4509817fbdce179d6470b"} Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.910400 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nrhl6" event={"ID":"f75a50cf-7dae-420b-a00d-671ede3cb6f7","Type":"ContainerDied","Data":"d0015edf4c699a29515d340ae2c01786195a0e9d51c4459eae81be82178c452a"} Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.910459 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nrhl6" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.912810 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tbqzs" event={"ID":"1e3d2cd4-4311-4454-b257-20a3caf243b3","Type":"ContainerStarted","Data":"313c18eb1c126858c938cbb0519630cfedb0f11cc4071cb91007fa8807d62753"} Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.912849 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tbqzs" event={"ID":"1e3d2cd4-4311-4454-b257-20a3caf243b3","Type":"ContainerStarted","Data":"60097ca12a5b208298fe3999bd1e665bf320413d781422c5a1f9dc265d840874"} Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.913773 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-tbqzs" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.920137 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c9z42"] Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.920187 4948 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-tbqzs container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.57:8080/healthz\": dial tcp 10.217.0.57:8080: connect: connection refused" start-of-body= Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.920267 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-tbqzs" podUID="1e3d2cd4-4311-4454-b257-20a3caf243b3" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.57:8080/healthz\": dial tcp 10.217.0.57:8080: connect: connection refused" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.920850 4948 scope.go:117] "RemoveContainer" containerID="bfea6ab4eabff6af6e0a3b30d304544693e06a2189af921dbd30d2b97a616f83" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.935414 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-c9z42"] Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.939130 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sjv8t"] Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.941921 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sjv8t"] Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.953532 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-tbqzs" podStartSLOduration=0.953514816 podStartE2EDuration="953.514816ms" podCreationTimestamp="2026-02-20 08:10:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:10:39.951504847 +0000 UTC m=+288.925999677" watchObservedRunningTime="2026-02-20 08:10:39.953514816 +0000 UTC m=+288.928009646" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.963187 4948 scope.go:117] "RemoveContainer" containerID="1d07ae65c7f4887c7cc6c699e956a80f175ba6deadc7da37863658727e044dbe" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.971762 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n8d7m"] Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.977914 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-n8d7m"] Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.986734 4948 scope.go:117] "RemoveContainer" containerID="1c4c2fe97c46c10efdb2a7e794505136dc520212167c78787ccbac94170a83c0" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.987655 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-frznz"] Feb 20 08:10:39 crc kubenswrapper[4948]: E0220 08:10:39.988381 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c4c2fe97c46c10efdb2a7e794505136dc520212167c78787ccbac94170a83c0\": container with ID starting with 1c4c2fe97c46c10efdb2a7e794505136dc520212167c78787ccbac94170a83c0 not found: ID does not exist" containerID="1c4c2fe97c46c10efdb2a7e794505136dc520212167c78787ccbac94170a83c0" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.988443 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c4c2fe97c46c10efdb2a7e794505136dc520212167c78787ccbac94170a83c0"} err="failed to get container status \"1c4c2fe97c46c10efdb2a7e794505136dc520212167c78787ccbac94170a83c0\": rpc error: code = NotFound desc = could not find container \"1c4c2fe97c46c10efdb2a7e794505136dc520212167c78787ccbac94170a83c0\": container with ID starting with 1c4c2fe97c46c10efdb2a7e794505136dc520212167c78787ccbac94170a83c0 not found: ID does not exist" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.988475 4948 scope.go:117] "RemoveContainer" containerID="bfea6ab4eabff6af6e0a3b30d304544693e06a2189af921dbd30d2b97a616f83" Feb 20 08:10:39 crc kubenswrapper[4948]: E0220 08:10:39.989320 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfea6ab4eabff6af6e0a3b30d304544693e06a2189af921dbd30d2b97a616f83\": container with ID starting with bfea6ab4eabff6af6e0a3b30d304544693e06a2189af921dbd30d2b97a616f83 not found: ID does not exist" containerID="bfea6ab4eabff6af6e0a3b30d304544693e06a2189af921dbd30d2b97a616f83" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.989407 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfea6ab4eabff6af6e0a3b30d304544693e06a2189af921dbd30d2b97a616f83"} err="failed to get container status \"bfea6ab4eabff6af6e0a3b30d304544693e06a2189af921dbd30d2b97a616f83\": rpc error: code = NotFound desc = could not find container \"bfea6ab4eabff6af6e0a3b30d304544693e06a2189af921dbd30d2b97a616f83\": container with ID starting with bfea6ab4eabff6af6e0a3b30d304544693e06a2189af921dbd30d2b97a616f83 not found: ID does not exist" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.989492 4948 scope.go:117] "RemoveContainer" containerID="1d07ae65c7f4887c7cc6c699e956a80f175ba6deadc7da37863658727e044dbe" Feb 20 08:10:39 crc kubenswrapper[4948]: E0220 08:10:39.990158 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d07ae65c7f4887c7cc6c699e956a80f175ba6deadc7da37863658727e044dbe\": container with ID starting with 1d07ae65c7f4887c7cc6c699e956a80f175ba6deadc7da37863658727e044dbe not found: ID does not exist" containerID="1d07ae65c7f4887c7cc6c699e956a80f175ba6deadc7da37863658727e044dbe" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.990207 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d07ae65c7f4887c7cc6c699e956a80f175ba6deadc7da37863658727e044dbe"} err="failed to get container status \"1d07ae65c7f4887c7cc6c699e956a80f175ba6deadc7da37863658727e044dbe\": rpc error: code = NotFound desc = could not find container \"1d07ae65c7f4887c7cc6c699e956a80f175ba6deadc7da37863658727e044dbe\": container with ID starting with 1d07ae65c7f4887c7cc6c699e956a80f175ba6deadc7da37863658727e044dbe not found: ID does not exist" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.990240 4948 scope.go:117] "RemoveContainer" containerID="fb097a14ed0ab7d5af6abab587a1c01e68baf8b43ccad645f65796167b0c90bd" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.991778 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-frznz"] Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.000250 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nrhl6"] Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.004032 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nrhl6"] Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.011459 4948 scope.go:117] "RemoveContainer" containerID="27b14a9abd1358ef732430763a33e8727bba43c949149d1b6052fc586fb0bd00" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.025717 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.028781 4948 scope.go:117] "RemoveContainer" containerID="6e6f9609316227fbb42f161491bc563af5bbab43a8ee401a26665bf588b0e5d7" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.044215 4948 scope.go:117] "RemoveContainer" containerID="fb097a14ed0ab7d5af6abab587a1c01e68baf8b43ccad645f65796167b0c90bd" Feb 20 08:10:40 crc kubenswrapper[4948]: E0220 08:10:40.044819 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb097a14ed0ab7d5af6abab587a1c01e68baf8b43ccad645f65796167b0c90bd\": container with ID starting with fb097a14ed0ab7d5af6abab587a1c01e68baf8b43ccad645f65796167b0c90bd not found: ID does not exist" containerID="fb097a14ed0ab7d5af6abab587a1c01e68baf8b43ccad645f65796167b0c90bd" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.044945 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb097a14ed0ab7d5af6abab587a1c01e68baf8b43ccad645f65796167b0c90bd"} err="failed to get container status \"fb097a14ed0ab7d5af6abab587a1c01e68baf8b43ccad645f65796167b0c90bd\": rpc error: code = NotFound desc = could not find container \"fb097a14ed0ab7d5af6abab587a1c01e68baf8b43ccad645f65796167b0c90bd\": container with ID starting with fb097a14ed0ab7d5af6abab587a1c01e68baf8b43ccad645f65796167b0c90bd not found: ID does not exist" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.045056 4948 scope.go:117] "RemoveContainer" containerID="27b14a9abd1358ef732430763a33e8727bba43c949149d1b6052fc586fb0bd00" Feb 20 08:10:40 crc kubenswrapper[4948]: E0220 08:10:40.045485 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27b14a9abd1358ef732430763a33e8727bba43c949149d1b6052fc586fb0bd00\": container with ID starting with 27b14a9abd1358ef732430763a33e8727bba43c949149d1b6052fc586fb0bd00 not found: ID does not exist" containerID="27b14a9abd1358ef732430763a33e8727bba43c949149d1b6052fc586fb0bd00" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.045571 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27b14a9abd1358ef732430763a33e8727bba43c949149d1b6052fc586fb0bd00"} err="failed to get container status \"27b14a9abd1358ef732430763a33e8727bba43c949149d1b6052fc586fb0bd00\": rpc error: code = NotFound desc = could not find container \"27b14a9abd1358ef732430763a33e8727bba43c949149d1b6052fc586fb0bd00\": container with ID starting with 27b14a9abd1358ef732430763a33e8727bba43c949149d1b6052fc586fb0bd00 not found: ID does not exist" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.045651 4948 scope.go:117] "RemoveContainer" containerID="6e6f9609316227fbb42f161491bc563af5bbab43a8ee401a26665bf588b0e5d7" Feb 20 08:10:40 crc kubenswrapper[4948]: E0220 08:10:40.046103 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e6f9609316227fbb42f161491bc563af5bbab43a8ee401a26665bf588b0e5d7\": container with ID starting with 6e6f9609316227fbb42f161491bc563af5bbab43a8ee401a26665bf588b0e5d7 not found: ID does not exist" containerID="6e6f9609316227fbb42f161491bc563af5bbab43a8ee401a26665bf588b0e5d7" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.046206 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e6f9609316227fbb42f161491bc563af5bbab43a8ee401a26665bf588b0e5d7"} err="failed to get container status \"6e6f9609316227fbb42f161491bc563af5bbab43a8ee401a26665bf588b0e5d7\": rpc error: code = NotFound desc = could not find container \"6e6f9609316227fbb42f161491bc563af5bbab43a8ee401a26665bf588b0e5d7\": container with ID starting with 6e6f9609316227fbb42f161491bc563af5bbab43a8ee401a26665bf588b0e5d7 not found: ID does not exist" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.046286 4948 scope.go:117] "RemoveContainer" containerID="15687151f4ed59c06cf2071d3204504981cf4a0c6cdb40d25a5dbd083313ed65" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.061730 4948 scope.go:117] "RemoveContainer" containerID="15687151f4ed59c06cf2071d3204504981cf4a0c6cdb40d25a5dbd083313ed65" Feb 20 08:10:40 crc kubenswrapper[4948]: E0220 08:10:40.062723 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15687151f4ed59c06cf2071d3204504981cf4a0c6cdb40d25a5dbd083313ed65\": container with ID starting with 15687151f4ed59c06cf2071d3204504981cf4a0c6cdb40d25a5dbd083313ed65 not found: ID does not exist" containerID="15687151f4ed59c06cf2071d3204504981cf4a0c6cdb40d25a5dbd083313ed65" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.062762 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15687151f4ed59c06cf2071d3204504981cf4a0c6cdb40d25a5dbd083313ed65"} err="failed to get container status \"15687151f4ed59c06cf2071d3204504981cf4a0c6cdb40d25a5dbd083313ed65\": rpc error: code = NotFound desc = could not find container \"15687151f4ed59c06cf2071d3204504981cf4a0c6cdb40d25a5dbd083313ed65\": container with ID starting with 15687151f4ed59c06cf2071d3204504981cf4a0c6cdb40d25a5dbd083313ed65 not found: ID does not exist" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.062790 4948 scope.go:117] "RemoveContainer" containerID="acfb3ae1070bc662949f97e63b0d95f3a1957ddfef18898cde66fcbe8186d502" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.076981 4948 scope.go:117] "RemoveContainer" containerID="5e9a0379202da2a09189213d51042628595d46ea2ac1801d0bcbd926088826d7" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.094332 4948 scope.go:117] "RemoveContainer" containerID="8ae4d6ef7b64ba80a0da276ad7998550243e3a26aaa0bbf3282f22cf4366103f" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.105866 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.110917 4948 scope.go:117] "RemoveContainer" containerID="acfb3ae1070bc662949f97e63b0d95f3a1957ddfef18898cde66fcbe8186d502" Feb 20 08:10:40 crc kubenswrapper[4948]: E0220 08:10:40.111466 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acfb3ae1070bc662949f97e63b0d95f3a1957ddfef18898cde66fcbe8186d502\": container with ID starting with acfb3ae1070bc662949f97e63b0d95f3a1957ddfef18898cde66fcbe8186d502 not found: ID does not exist" containerID="acfb3ae1070bc662949f97e63b0d95f3a1957ddfef18898cde66fcbe8186d502" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.111579 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acfb3ae1070bc662949f97e63b0d95f3a1957ddfef18898cde66fcbe8186d502"} err="failed to get container status \"acfb3ae1070bc662949f97e63b0d95f3a1957ddfef18898cde66fcbe8186d502\": rpc error: code = NotFound desc = could not find container \"acfb3ae1070bc662949f97e63b0d95f3a1957ddfef18898cde66fcbe8186d502\": container with ID starting with acfb3ae1070bc662949f97e63b0d95f3a1957ddfef18898cde66fcbe8186d502 not found: ID does not exist" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.111677 4948 scope.go:117] "RemoveContainer" containerID="5e9a0379202da2a09189213d51042628595d46ea2ac1801d0bcbd926088826d7" Feb 20 08:10:40 crc kubenswrapper[4948]: E0220 08:10:40.112099 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e9a0379202da2a09189213d51042628595d46ea2ac1801d0bcbd926088826d7\": container with ID starting with 5e9a0379202da2a09189213d51042628595d46ea2ac1801d0bcbd926088826d7 not found: ID does not exist" containerID="5e9a0379202da2a09189213d51042628595d46ea2ac1801d0bcbd926088826d7" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.112206 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e9a0379202da2a09189213d51042628595d46ea2ac1801d0bcbd926088826d7"} err="failed to get container status \"5e9a0379202da2a09189213d51042628595d46ea2ac1801d0bcbd926088826d7\": rpc error: code = NotFound desc = could not find container \"5e9a0379202da2a09189213d51042628595d46ea2ac1801d0bcbd926088826d7\": container with ID starting with 5e9a0379202da2a09189213d51042628595d46ea2ac1801d0bcbd926088826d7 not found: ID does not exist" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.112303 4948 scope.go:117] "RemoveContainer" containerID="8ae4d6ef7b64ba80a0da276ad7998550243e3a26aaa0bbf3282f22cf4366103f" Feb 20 08:10:40 crc kubenswrapper[4948]: E0220 08:10:40.114046 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ae4d6ef7b64ba80a0da276ad7998550243e3a26aaa0bbf3282f22cf4366103f\": container with ID starting with 8ae4d6ef7b64ba80a0da276ad7998550243e3a26aaa0bbf3282f22cf4366103f not found: ID does not exist" containerID="8ae4d6ef7b64ba80a0da276ad7998550243e3a26aaa0bbf3282f22cf4366103f" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.114149 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ae4d6ef7b64ba80a0da276ad7998550243e3a26aaa0bbf3282f22cf4366103f"} err="failed to get container status \"8ae4d6ef7b64ba80a0da276ad7998550243e3a26aaa0bbf3282f22cf4366103f\": rpc error: code = NotFound desc = could not find container \"8ae4d6ef7b64ba80a0da276ad7998550243e3a26aaa0bbf3282f22cf4366103f\": container with ID starting with 8ae4d6ef7b64ba80a0da276ad7998550243e3a26aaa0bbf3282f22cf4366103f not found: ID does not exist" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.114249 4948 scope.go:117] "RemoveContainer" containerID="51c14bb6285644fd84e90599149f7f330a4395b58ff4509817fbdce179d6470b" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.133522 4948 scope.go:117] "RemoveContainer" containerID="ed76ff8cc753dd2cd048a3406074573b902071e966e21613ec37f782fb620d37" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.150603 4948 scope.go:117] "RemoveContainer" containerID="df3ff28a9e161bcd3a383de17aefbe231a72f4c519e331260796574dc38d393e" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.152571 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.170268 4948 scope.go:117] "RemoveContainer" containerID="51c14bb6285644fd84e90599149f7f330a4395b58ff4509817fbdce179d6470b" Feb 20 08:10:40 crc kubenswrapper[4948]: E0220 08:10:40.170871 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51c14bb6285644fd84e90599149f7f330a4395b58ff4509817fbdce179d6470b\": container with ID starting with 51c14bb6285644fd84e90599149f7f330a4395b58ff4509817fbdce179d6470b not found: ID does not exist" containerID="51c14bb6285644fd84e90599149f7f330a4395b58ff4509817fbdce179d6470b" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.170944 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51c14bb6285644fd84e90599149f7f330a4395b58ff4509817fbdce179d6470b"} err="failed to get container status \"51c14bb6285644fd84e90599149f7f330a4395b58ff4509817fbdce179d6470b\": rpc error: code = NotFound desc = could not find container \"51c14bb6285644fd84e90599149f7f330a4395b58ff4509817fbdce179d6470b\": container with ID starting with 51c14bb6285644fd84e90599149f7f330a4395b58ff4509817fbdce179d6470b not found: ID does not exist" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.171013 4948 scope.go:117] "RemoveContainer" containerID="ed76ff8cc753dd2cd048a3406074573b902071e966e21613ec37f782fb620d37" Feb 20 08:10:40 crc kubenswrapper[4948]: E0220 08:10:40.171957 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed76ff8cc753dd2cd048a3406074573b902071e966e21613ec37f782fb620d37\": container with ID starting with ed76ff8cc753dd2cd048a3406074573b902071e966e21613ec37f782fb620d37 not found: ID does not exist" containerID="ed76ff8cc753dd2cd048a3406074573b902071e966e21613ec37f782fb620d37" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.172022 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed76ff8cc753dd2cd048a3406074573b902071e966e21613ec37f782fb620d37"} err="failed to get container status \"ed76ff8cc753dd2cd048a3406074573b902071e966e21613ec37f782fb620d37\": rpc error: code = NotFound desc = could not find container \"ed76ff8cc753dd2cd048a3406074573b902071e966e21613ec37f782fb620d37\": container with ID starting with ed76ff8cc753dd2cd048a3406074573b902071e966e21613ec37f782fb620d37 not found: ID does not exist" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.172058 4948 scope.go:117] "RemoveContainer" containerID="df3ff28a9e161bcd3a383de17aefbe231a72f4c519e331260796574dc38d393e" Feb 20 08:10:40 crc kubenswrapper[4948]: E0220 08:10:40.172485 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df3ff28a9e161bcd3a383de17aefbe231a72f4c519e331260796574dc38d393e\": container with ID starting with df3ff28a9e161bcd3a383de17aefbe231a72f4c519e331260796574dc38d393e not found: ID does not exist" containerID="df3ff28a9e161bcd3a383de17aefbe231a72f4c519e331260796574dc38d393e" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.172593 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df3ff28a9e161bcd3a383de17aefbe231a72f4c519e331260796574dc38d393e"} err="failed to get container status \"df3ff28a9e161bcd3a383de17aefbe231a72f4c519e331260796574dc38d393e\": rpc error: code = NotFound desc = could not find container \"df3ff28a9e161bcd3a383de17aefbe231a72f4c519e331260796574dc38d393e\": container with ID starting with df3ff28a9e161bcd3a383de17aefbe231a72f4c519e331260796574dc38d393e not found: ID does not exist" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.174268 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.334873 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.349746 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.397873 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.425224 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.460868 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.584851 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.719298 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.778283 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.885936 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.904205 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.905322 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.945049 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-tbqzs" Feb 20 08:10:41 crc kubenswrapper[4948]: I0220 08:10:41.264455 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 20 08:10:41 crc kubenswrapper[4948]: I0220 08:10:41.331441 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 20 08:10:41 crc kubenswrapper[4948]: I0220 08:10:41.384908 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 20 08:10:41 crc kubenswrapper[4948]: I0220 08:10:41.514375 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 20 08:10:41 crc kubenswrapper[4948]: I0220 08:10:41.734810 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00f4b4c1-af36-4ccd-a3d1-dd6df1186338" path="/var/lib/kubelet/pods/00f4b4c1-af36-4ccd-a3d1-dd6df1186338/volumes" Feb 20 08:10:41 crc kubenswrapper[4948]: I0220 08:10:41.736222 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95a270e5-7a08-4238-a65d-96986f2d2c39" path="/var/lib/kubelet/pods/95a270e5-7a08-4238-a65d-96986f2d2c39/volumes" Feb 20 08:10:41 crc kubenswrapper[4948]: I0220 08:10:41.738184 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c53ddd33-0f9f-4794-b346-7d48a6c09c9b" path="/var/lib/kubelet/pods/c53ddd33-0f9f-4794-b346-7d48a6c09c9b/volumes" Feb 20 08:10:41 crc kubenswrapper[4948]: I0220 08:10:41.740316 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" path="/var/lib/kubelet/pods/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9/volumes" Feb 20 08:10:41 crc kubenswrapper[4948]: I0220 08:10:41.741620 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f75a50cf-7dae-420b-a00d-671ede3cb6f7" path="/var/lib/kubelet/pods/f75a50cf-7dae-420b-a00d-671ede3cb6f7/volumes" Feb 20 08:10:41 crc kubenswrapper[4948]: I0220 08:10:41.849828 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 20 08:10:41 crc kubenswrapper[4948]: I0220 08:10:41.879837 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 20 08:10:41 crc kubenswrapper[4948]: I0220 08:10:41.978863 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 20 08:10:41 crc kubenswrapper[4948]: I0220 08:10:41.990024 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 20 08:10:42 crc kubenswrapper[4948]: I0220 08:10:42.935709 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 20 08:10:42 crc kubenswrapper[4948]: I0220 08:10:42.944953 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 20 08:10:42 crc kubenswrapper[4948]: I0220 08:10:42.945110 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:10:42 crc kubenswrapper[4948]: I0220 08:10:42.980669 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 20 08:10:42 crc kubenswrapper[4948]: I0220 08:10:42.980740 4948 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="955d8cda21a3da3121dfa54a2504d8b0f8470dbb104720ff4e180c5ad7329f47" exitCode=137 Feb 20 08:10:42 crc kubenswrapper[4948]: I0220 08:10:42.981225 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:10:42 crc kubenswrapper[4948]: I0220 08:10:42.981419 4948 scope.go:117] "RemoveContainer" containerID="955d8cda21a3da3121dfa54a2504d8b0f8470dbb104720ff4e180c5ad7329f47" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.004517 4948 scope.go:117] "RemoveContainer" containerID="955d8cda21a3da3121dfa54a2504d8b0f8470dbb104720ff4e180c5ad7329f47" Feb 20 08:10:43 crc kubenswrapper[4948]: E0220 08:10:43.005792 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"955d8cda21a3da3121dfa54a2504d8b0f8470dbb104720ff4e180c5ad7329f47\": container with ID starting with 955d8cda21a3da3121dfa54a2504d8b0f8470dbb104720ff4e180c5ad7329f47 not found: ID does not exist" containerID="955d8cda21a3da3121dfa54a2504d8b0f8470dbb104720ff4e180c5ad7329f47" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.005851 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"955d8cda21a3da3121dfa54a2504d8b0f8470dbb104720ff4e180c5ad7329f47"} err="failed to get container status \"955d8cda21a3da3121dfa54a2504d8b0f8470dbb104720ff4e180c5ad7329f47\": rpc error: code = NotFound desc = could not find container \"955d8cda21a3da3121dfa54a2504d8b0f8470dbb104720ff4e180c5ad7329f47\": container with ID starting with 955d8cda21a3da3121dfa54a2504d8b0f8470dbb104720ff4e180c5ad7329f47 not found: ID does not exist" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.079705 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.079744 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.079823 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.079881 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.079905 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.079931 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.079937 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.079989 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.080000 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.080337 4948 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.080358 4948 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.080372 4948 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.080381 4948 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.086895 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.158455 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.181814 4948 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.735571 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.736311 4948 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.748915 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.748968 4948 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="872f9266-e8c5-46f8-bddb-c112fa6260a9" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.753326 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.753371 4948 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="872f9266-e8c5-46f8-bddb-c112fa6260a9" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.880698 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 20 08:10:51 crc kubenswrapper[4948]: I0220 08:10:51.489820 4948 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.474152 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dl7nj"] Feb 20 08:11:05 crc kubenswrapper[4948]: E0220 08:11:05.475029 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95a270e5-7a08-4238-a65d-96986f2d2c39" containerName="marketplace-operator" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475050 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="95a270e5-7a08-4238-a65d-96986f2d2c39" containerName="marketplace-operator" Feb 20 08:11:05 crc kubenswrapper[4948]: E0220 08:11:05.475073 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f75a50cf-7dae-420b-a00d-671ede3cb6f7" containerName="extract-content" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475086 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f75a50cf-7dae-420b-a00d-671ede3cb6f7" containerName="extract-content" Feb 20 08:11:05 crc kubenswrapper[4948]: E0220 08:11:05.475103 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00f4b4c1-af36-4ccd-a3d1-dd6df1186338" containerName="extract-content" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475115 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="00f4b4c1-af36-4ccd-a3d1-dd6df1186338" containerName="extract-content" Feb 20 08:11:05 crc kubenswrapper[4948]: E0220 08:11:05.475130 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" containerName="registry-server" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475141 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" containerName="registry-server" Feb 20 08:11:05 crc kubenswrapper[4948]: E0220 08:11:05.475161 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f75a50cf-7dae-420b-a00d-671ede3cb6f7" containerName="registry-server" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475172 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f75a50cf-7dae-420b-a00d-671ede3cb6f7" containerName="registry-server" Feb 20 08:11:05 crc kubenswrapper[4948]: E0220 08:11:05.475190 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f75a50cf-7dae-420b-a00d-671ede3cb6f7" containerName="extract-utilities" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475203 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f75a50cf-7dae-420b-a00d-671ede3cb6f7" containerName="extract-utilities" Feb 20 08:11:05 crc kubenswrapper[4948]: E0220 08:11:05.475220 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" containerName="extract-utilities" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475232 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" containerName="extract-utilities" Feb 20 08:11:05 crc kubenswrapper[4948]: E0220 08:11:05.475251 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c53ddd33-0f9f-4794-b346-7d48a6c09c9b" containerName="registry-server" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475263 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="c53ddd33-0f9f-4794-b346-7d48a6c09c9b" containerName="registry-server" Feb 20 08:11:05 crc kubenswrapper[4948]: E0220 08:11:05.475276 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00f4b4c1-af36-4ccd-a3d1-dd6df1186338" containerName="extract-utilities" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475289 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="00f4b4c1-af36-4ccd-a3d1-dd6df1186338" containerName="extract-utilities" Feb 20 08:11:05 crc kubenswrapper[4948]: E0220 08:11:05.475305 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" containerName="extract-content" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475316 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" containerName="extract-content" Feb 20 08:11:05 crc kubenswrapper[4948]: E0220 08:11:05.475334 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c53ddd33-0f9f-4794-b346-7d48a6c09c9b" containerName="extract-content" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475346 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="c53ddd33-0f9f-4794-b346-7d48a6c09c9b" containerName="extract-content" Feb 20 08:11:05 crc kubenswrapper[4948]: E0220 08:11:05.475361 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00f4b4c1-af36-4ccd-a3d1-dd6df1186338" containerName="registry-server" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475373 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="00f4b4c1-af36-4ccd-a3d1-dd6df1186338" containerName="registry-server" Feb 20 08:11:05 crc kubenswrapper[4948]: E0220 08:11:05.475394 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c53ddd33-0f9f-4794-b346-7d48a6c09c9b" containerName="extract-utilities" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475405 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="c53ddd33-0f9f-4794-b346-7d48a6c09c9b" containerName="extract-utilities" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475561 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="00f4b4c1-af36-4ccd-a3d1-dd6df1186338" containerName="registry-server" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475581 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f75a50cf-7dae-420b-a00d-671ede3cb6f7" containerName="registry-server" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475595 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" containerName="registry-server" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475613 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="95a270e5-7a08-4238-a65d-96986f2d2c39" containerName="marketplace-operator" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475630 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="c53ddd33-0f9f-4794-b346-7d48a6c09c9b" containerName="registry-server" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.476852 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dl7nj" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.481192 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.489621 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e759dc6-36e8-4a97-b994-e657f1b5d540-utilities\") pod \"community-operators-dl7nj\" (UID: \"0e759dc6-36e8-4a97-b994-e657f1b5d540\") " pod="openshift-marketplace/community-operators-dl7nj" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.489721 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqd6w\" (UniqueName: \"kubernetes.io/projected/0e759dc6-36e8-4a97-b994-e657f1b5d540-kube-api-access-lqd6w\") pod \"community-operators-dl7nj\" (UID: \"0e759dc6-36e8-4a97-b994-e657f1b5d540\") " pod="openshift-marketplace/community-operators-dl7nj" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.489842 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e759dc6-36e8-4a97-b994-e657f1b5d540-catalog-content\") pod \"community-operators-dl7nj\" (UID: \"0e759dc6-36e8-4a97-b994-e657f1b5d540\") " pod="openshift-marketplace/community-operators-dl7nj" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.496112 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dl7nj"] Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.591201 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e759dc6-36e8-4a97-b994-e657f1b5d540-catalog-content\") pod \"community-operators-dl7nj\" (UID: \"0e759dc6-36e8-4a97-b994-e657f1b5d540\") " pod="openshift-marketplace/community-operators-dl7nj" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.591403 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e759dc6-36e8-4a97-b994-e657f1b5d540-utilities\") pod \"community-operators-dl7nj\" (UID: \"0e759dc6-36e8-4a97-b994-e657f1b5d540\") " pod="openshift-marketplace/community-operators-dl7nj" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.591439 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqd6w\" (UniqueName: \"kubernetes.io/projected/0e759dc6-36e8-4a97-b994-e657f1b5d540-kube-api-access-lqd6w\") pod \"community-operators-dl7nj\" (UID: \"0e759dc6-36e8-4a97-b994-e657f1b5d540\") " pod="openshift-marketplace/community-operators-dl7nj" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.592825 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e759dc6-36e8-4a97-b994-e657f1b5d540-utilities\") pod \"community-operators-dl7nj\" (UID: \"0e759dc6-36e8-4a97-b994-e657f1b5d540\") " pod="openshift-marketplace/community-operators-dl7nj" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.593014 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e759dc6-36e8-4a97-b994-e657f1b5d540-catalog-content\") pod \"community-operators-dl7nj\" (UID: \"0e759dc6-36e8-4a97-b994-e657f1b5d540\") " pod="openshift-marketplace/community-operators-dl7nj" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.618160 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqd6w\" (UniqueName: \"kubernetes.io/projected/0e759dc6-36e8-4a97-b994-e657f1b5d540-kube-api-access-lqd6w\") pod \"community-operators-dl7nj\" (UID: \"0e759dc6-36e8-4a97-b994-e657f1b5d540\") " pod="openshift-marketplace/community-operators-dl7nj" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.664295 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tt6t6"] Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.666334 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tt6t6" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.669304 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.679754 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tt6t6"] Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.692515 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsp6z\" (UniqueName: \"kubernetes.io/projected/94e3bb0b-818a-491e-8e1e-a1664f3a9f12-kube-api-access-nsp6z\") pod \"certified-operators-tt6t6\" (UID: \"94e3bb0b-818a-491e-8e1e-a1664f3a9f12\") " pod="openshift-marketplace/certified-operators-tt6t6" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.692568 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94e3bb0b-818a-491e-8e1e-a1664f3a9f12-catalog-content\") pod \"certified-operators-tt6t6\" (UID: \"94e3bb0b-818a-491e-8e1e-a1664f3a9f12\") " pod="openshift-marketplace/certified-operators-tt6t6" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.692633 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94e3bb0b-818a-491e-8e1e-a1664f3a9f12-utilities\") pod \"certified-operators-tt6t6\" (UID: \"94e3bb0b-818a-491e-8e1e-a1664f3a9f12\") " pod="openshift-marketplace/certified-operators-tt6t6" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.794408 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsp6z\" (UniqueName: \"kubernetes.io/projected/94e3bb0b-818a-491e-8e1e-a1664f3a9f12-kube-api-access-nsp6z\") pod \"certified-operators-tt6t6\" (UID: \"94e3bb0b-818a-491e-8e1e-a1664f3a9f12\") " pod="openshift-marketplace/certified-operators-tt6t6" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.794537 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94e3bb0b-818a-491e-8e1e-a1664f3a9f12-catalog-content\") pod \"certified-operators-tt6t6\" (UID: \"94e3bb0b-818a-491e-8e1e-a1664f3a9f12\") " pod="openshift-marketplace/certified-operators-tt6t6" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.794734 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94e3bb0b-818a-491e-8e1e-a1664f3a9f12-utilities\") pod \"certified-operators-tt6t6\" (UID: \"94e3bb0b-818a-491e-8e1e-a1664f3a9f12\") " pod="openshift-marketplace/certified-operators-tt6t6" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.795474 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94e3bb0b-818a-491e-8e1e-a1664f3a9f12-catalog-content\") pod \"certified-operators-tt6t6\" (UID: \"94e3bb0b-818a-491e-8e1e-a1664f3a9f12\") " pod="openshift-marketplace/certified-operators-tt6t6" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.795890 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94e3bb0b-818a-491e-8e1e-a1664f3a9f12-utilities\") pod \"certified-operators-tt6t6\" (UID: \"94e3bb0b-818a-491e-8e1e-a1664f3a9f12\") " pod="openshift-marketplace/certified-operators-tt6t6" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.819886 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsp6z\" (UniqueName: \"kubernetes.io/projected/94e3bb0b-818a-491e-8e1e-a1664f3a9f12-kube-api-access-nsp6z\") pod \"certified-operators-tt6t6\" (UID: \"94e3bb0b-818a-491e-8e1e-a1664f3a9f12\") " pod="openshift-marketplace/certified-operators-tt6t6" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.844666 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dl7nj" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.989500 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tt6t6" Feb 20 08:11:06 crc kubenswrapper[4948]: I0220 08:11:06.104921 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dl7nj"] Feb 20 08:11:06 crc kubenswrapper[4948]: I0220 08:11:06.122960 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dl7nj" event={"ID":"0e759dc6-36e8-4a97-b994-e657f1b5d540","Type":"ContainerStarted","Data":"bf3ed40d3cc7b0698c95b69e2698db264f0323aa1832e449fcfe404052daf599"} Feb 20 08:11:06 crc kubenswrapper[4948]: I0220 08:11:06.249219 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tt6t6"] Feb 20 08:11:06 crc kubenswrapper[4948]: W0220 08:11:06.255704 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod94e3bb0b_818a_491e_8e1e_a1664f3a9f12.slice/crio-a46eccf52e321b9645691299b6786a0989ece4633f0c772e19e8616200646554 WatchSource:0}: Error finding container a46eccf52e321b9645691299b6786a0989ece4633f0c772e19e8616200646554: Status 404 returned error can't find the container with id a46eccf52e321b9645691299b6786a0989ece4633f0c772e19e8616200646554 Feb 20 08:11:07 crc kubenswrapper[4948]: I0220 08:11:07.132703 4948 generic.go:334] "Generic (PLEG): container finished" podID="0e759dc6-36e8-4a97-b994-e657f1b5d540" containerID="7a7848ac580be195cccad9466f97fab3d16ed5cce26ab1658187f2cafddd55eb" exitCode=0 Feb 20 08:11:07 crc kubenswrapper[4948]: I0220 08:11:07.132835 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dl7nj" event={"ID":"0e759dc6-36e8-4a97-b994-e657f1b5d540","Type":"ContainerDied","Data":"7a7848ac580be195cccad9466f97fab3d16ed5cce26ab1658187f2cafddd55eb"} Feb 20 08:11:07 crc kubenswrapper[4948]: I0220 08:11:07.137267 4948 generic.go:334] "Generic (PLEG): container finished" podID="94e3bb0b-818a-491e-8e1e-a1664f3a9f12" containerID="5404601cc08a73a981959eb4e9f568980c021e72d4fb5ec5a3c470eab642d7d1" exitCode=0 Feb 20 08:11:07 crc kubenswrapper[4948]: I0220 08:11:07.137354 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tt6t6" event={"ID":"94e3bb0b-818a-491e-8e1e-a1664f3a9f12","Type":"ContainerDied","Data":"5404601cc08a73a981959eb4e9f568980c021e72d4fb5ec5a3c470eab642d7d1"} Feb 20 08:11:07 crc kubenswrapper[4948]: I0220 08:11:07.137409 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tt6t6" event={"ID":"94e3bb0b-818a-491e-8e1e-a1664f3a9f12","Type":"ContainerStarted","Data":"a46eccf52e321b9645691299b6786a0989ece4633f0c772e19e8616200646554"} Feb 20 08:11:07 crc kubenswrapper[4948]: I0220 08:11:07.869709 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-x7bcm"] Feb 20 08:11:07 crc kubenswrapper[4948]: I0220 08:11:07.871030 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x7bcm" Feb 20 08:11:07 crc kubenswrapper[4948]: I0220 08:11:07.874800 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 20 08:11:07 crc kubenswrapper[4948]: I0220 08:11:07.890816 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x7bcm"] Feb 20 08:11:07 crc kubenswrapper[4948]: I0220 08:11:07.934326 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ec98112-e189-4a27-a6dd-8cb035596810-utilities\") pod \"redhat-marketplace-x7bcm\" (UID: \"6ec98112-e189-4a27-a6dd-8cb035596810\") " pod="openshift-marketplace/redhat-marketplace-x7bcm" Feb 20 08:11:07 crc kubenswrapper[4948]: I0220 08:11:07.934740 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ec98112-e189-4a27-a6dd-8cb035596810-catalog-content\") pod \"redhat-marketplace-x7bcm\" (UID: \"6ec98112-e189-4a27-a6dd-8cb035596810\") " pod="openshift-marketplace/redhat-marketplace-x7bcm" Feb 20 08:11:07 crc kubenswrapper[4948]: I0220 08:11:07.934764 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bplnw\" (UniqueName: \"kubernetes.io/projected/6ec98112-e189-4a27-a6dd-8cb035596810-kube-api-access-bplnw\") pod \"redhat-marketplace-x7bcm\" (UID: \"6ec98112-e189-4a27-a6dd-8cb035596810\") " pod="openshift-marketplace/redhat-marketplace-x7bcm" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.035641 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ec98112-e189-4a27-a6dd-8cb035596810-utilities\") pod \"redhat-marketplace-x7bcm\" (UID: \"6ec98112-e189-4a27-a6dd-8cb035596810\") " pod="openshift-marketplace/redhat-marketplace-x7bcm" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.035756 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ec98112-e189-4a27-a6dd-8cb035596810-catalog-content\") pod \"redhat-marketplace-x7bcm\" (UID: \"6ec98112-e189-4a27-a6dd-8cb035596810\") " pod="openshift-marketplace/redhat-marketplace-x7bcm" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.035792 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bplnw\" (UniqueName: \"kubernetes.io/projected/6ec98112-e189-4a27-a6dd-8cb035596810-kube-api-access-bplnw\") pod \"redhat-marketplace-x7bcm\" (UID: \"6ec98112-e189-4a27-a6dd-8cb035596810\") " pod="openshift-marketplace/redhat-marketplace-x7bcm" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.036153 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ec98112-e189-4a27-a6dd-8cb035596810-utilities\") pod \"redhat-marketplace-x7bcm\" (UID: \"6ec98112-e189-4a27-a6dd-8cb035596810\") " pod="openshift-marketplace/redhat-marketplace-x7bcm" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.036416 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ec98112-e189-4a27-a6dd-8cb035596810-catalog-content\") pod \"redhat-marketplace-x7bcm\" (UID: \"6ec98112-e189-4a27-a6dd-8cb035596810\") " pod="openshift-marketplace/redhat-marketplace-x7bcm" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.071043 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jbtm5"] Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.073320 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jbtm5" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.076931 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.082645 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bplnw\" (UniqueName: \"kubernetes.io/projected/6ec98112-e189-4a27-a6dd-8cb035596810-kube-api-access-bplnw\") pod \"redhat-marketplace-x7bcm\" (UID: \"6ec98112-e189-4a27-a6dd-8cb035596810\") " pod="openshift-marketplace/redhat-marketplace-x7bcm" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.084864 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jbtm5"] Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.136463 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zmfg\" (UniqueName: \"kubernetes.io/projected/5ee242c9-9783-44cb-a4dc-1ea50a7ae185-kube-api-access-8zmfg\") pod \"redhat-operators-jbtm5\" (UID: \"5ee242c9-9783-44cb-a4dc-1ea50a7ae185\") " pod="openshift-marketplace/redhat-operators-jbtm5" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.136517 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ee242c9-9783-44cb-a4dc-1ea50a7ae185-catalog-content\") pod \"redhat-operators-jbtm5\" (UID: \"5ee242c9-9783-44cb-a4dc-1ea50a7ae185\") " pod="openshift-marketplace/redhat-operators-jbtm5" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.136722 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ee242c9-9783-44cb-a4dc-1ea50a7ae185-utilities\") pod \"redhat-operators-jbtm5\" (UID: \"5ee242c9-9783-44cb-a4dc-1ea50a7ae185\") " pod="openshift-marketplace/redhat-operators-jbtm5" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.151942 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tt6t6" event={"ID":"94e3bb0b-818a-491e-8e1e-a1664f3a9f12","Type":"ContainerStarted","Data":"f1888dbac65f5a2dda53ac9b186d1b21ab5834fdf1cf175e2d58d53cb700d38b"} Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.216364 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x7bcm" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.238146 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zmfg\" (UniqueName: \"kubernetes.io/projected/5ee242c9-9783-44cb-a4dc-1ea50a7ae185-kube-api-access-8zmfg\") pod \"redhat-operators-jbtm5\" (UID: \"5ee242c9-9783-44cb-a4dc-1ea50a7ae185\") " pod="openshift-marketplace/redhat-operators-jbtm5" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.238204 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ee242c9-9783-44cb-a4dc-1ea50a7ae185-catalog-content\") pod \"redhat-operators-jbtm5\" (UID: \"5ee242c9-9783-44cb-a4dc-1ea50a7ae185\") " pod="openshift-marketplace/redhat-operators-jbtm5" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.238253 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ee242c9-9783-44cb-a4dc-1ea50a7ae185-utilities\") pod \"redhat-operators-jbtm5\" (UID: \"5ee242c9-9783-44cb-a4dc-1ea50a7ae185\") " pod="openshift-marketplace/redhat-operators-jbtm5" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.238872 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ee242c9-9783-44cb-a4dc-1ea50a7ae185-utilities\") pod \"redhat-operators-jbtm5\" (UID: \"5ee242c9-9783-44cb-a4dc-1ea50a7ae185\") " pod="openshift-marketplace/redhat-operators-jbtm5" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.239766 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ee242c9-9783-44cb-a4dc-1ea50a7ae185-catalog-content\") pod \"redhat-operators-jbtm5\" (UID: \"5ee242c9-9783-44cb-a4dc-1ea50a7ae185\") " pod="openshift-marketplace/redhat-operators-jbtm5" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.267711 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zmfg\" (UniqueName: \"kubernetes.io/projected/5ee242c9-9783-44cb-a4dc-1ea50a7ae185-kube-api-access-8zmfg\") pod \"redhat-operators-jbtm5\" (UID: \"5ee242c9-9783-44cb-a4dc-1ea50a7ae185\") " pod="openshift-marketplace/redhat-operators-jbtm5" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.464263 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x7bcm"] Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.466657 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jbtm5" Feb 20 08:11:08 crc kubenswrapper[4948]: W0220 08:11:08.551760 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ec98112_e189_4a27_a6dd_8cb035596810.slice/crio-3981f65f9b36579b160b94eda6a0c154957916c3d4d7e6b5d32e34517e7ed66c WatchSource:0}: Error finding container 3981f65f9b36579b160b94eda6a0c154957916c3d4d7e6b5d32e34517e7ed66c: Status 404 returned error can't find the container with id 3981f65f9b36579b160b94eda6a0c154957916c3d4d7e6b5d32e34517e7ed66c Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.696381 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jbtm5"] Feb 20 08:11:09 crc kubenswrapper[4948]: I0220 08:11:09.163297 4948 generic.go:334] "Generic (PLEG): container finished" podID="6ec98112-e189-4a27-a6dd-8cb035596810" containerID="b53992b5ed6f1bf4c2505a8d4b9914069c430e74cf847c56c06e70e9587d3fee" exitCode=0 Feb 20 08:11:09 crc kubenswrapper[4948]: I0220 08:11:09.163377 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x7bcm" event={"ID":"6ec98112-e189-4a27-a6dd-8cb035596810","Type":"ContainerDied","Data":"b53992b5ed6f1bf4c2505a8d4b9914069c430e74cf847c56c06e70e9587d3fee"} Feb 20 08:11:09 crc kubenswrapper[4948]: I0220 08:11:09.163684 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x7bcm" event={"ID":"6ec98112-e189-4a27-a6dd-8cb035596810","Type":"ContainerStarted","Data":"3981f65f9b36579b160b94eda6a0c154957916c3d4d7e6b5d32e34517e7ed66c"} Feb 20 08:11:09 crc kubenswrapper[4948]: I0220 08:11:09.167111 4948 generic.go:334] "Generic (PLEG): container finished" podID="0e759dc6-36e8-4a97-b994-e657f1b5d540" containerID="1564a74185083f2b92c7307fe86058881bf28dadd08bdfcd91b901b9737c79c3" exitCode=0 Feb 20 08:11:09 crc kubenswrapper[4948]: I0220 08:11:09.167241 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dl7nj" event={"ID":"0e759dc6-36e8-4a97-b994-e657f1b5d540","Type":"ContainerDied","Data":"1564a74185083f2b92c7307fe86058881bf28dadd08bdfcd91b901b9737c79c3"} Feb 20 08:11:09 crc kubenswrapper[4948]: I0220 08:11:09.170783 4948 generic.go:334] "Generic (PLEG): container finished" podID="94e3bb0b-818a-491e-8e1e-a1664f3a9f12" containerID="f1888dbac65f5a2dda53ac9b186d1b21ab5834fdf1cf175e2d58d53cb700d38b" exitCode=0 Feb 20 08:11:09 crc kubenswrapper[4948]: I0220 08:11:09.170876 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tt6t6" event={"ID":"94e3bb0b-818a-491e-8e1e-a1664f3a9f12","Type":"ContainerDied","Data":"f1888dbac65f5a2dda53ac9b186d1b21ab5834fdf1cf175e2d58d53cb700d38b"} Feb 20 08:11:09 crc kubenswrapper[4948]: I0220 08:11:09.173544 4948 generic.go:334] "Generic (PLEG): container finished" podID="5ee242c9-9783-44cb-a4dc-1ea50a7ae185" containerID="2218bda4dd124f74ba9c850581183328566b25c77cae393e864df3ccbbbabb4b" exitCode=0 Feb 20 08:11:09 crc kubenswrapper[4948]: I0220 08:11:09.173598 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbtm5" event={"ID":"5ee242c9-9783-44cb-a4dc-1ea50a7ae185","Type":"ContainerDied","Data":"2218bda4dd124f74ba9c850581183328566b25c77cae393e864df3ccbbbabb4b"} Feb 20 08:11:09 crc kubenswrapper[4948]: I0220 08:11:09.173629 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbtm5" event={"ID":"5ee242c9-9783-44cb-a4dc-1ea50a7ae185","Type":"ContainerStarted","Data":"dc93ca20aabfe84b972cccb2fcd744cdb44482b26917eb4a96821537d451ddf2"} Feb 20 08:11:10 crc kubenswrapper[4948]: I0220 08:11:10.180998 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbtm5" event={"ID":"5ee242c9-9783-44cb-a4dc-1ea50a7ae185","Type":"ContainerStarted","Data":"1da87f05a71242abf8441f6ff99031f27ba7460a91f1265f6101e951bc82ac76"} Feb 20 08:11:10 crc kubenswrapper[4948]: I0220 08:11:10.182850 4948 generic.go:334] "Generic (PLEG): container finished" podID="6ec98112-e189-4a27-a6dd-8cb035596810" containerID="d7bab80a1ea22ca8894e254a51e0da2b70189f46ceebd8ec1a2d3b1dbba46468" exitCode=0 Feb 20 08:11:10 crc kubenswrapper[4948]: I0220 08:11:10.183517 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x7bcm" event={"ID":"6ec98112-e189-4a27-a6dd-8cb035596810","Type":"ContainerDied","Data":"d7bab80a1ea22ca8894e254a51e0da2b70189f46ceebd8ec1a2d3b1dbba46468"} Feb 20 08:11:10 crc kubenswrapper[4948]: I0220 08:11:10.186707 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dl7nj" event={"ID":"0e759dc6-36e8-4a97-b994-e657f1b5d540","Type":"ContainerStarted","Data":"c2081b84a51e60cc8445891129a696d808670e42dc13946e716842b6c395bbad"} Feb 20 08:11:10 crc kubenswrapper[4948]: I0220 08:11:10.189006 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tt6t6" event={"ID":"94e3bb0b-818a-491e-8e1e-a1664f3a9f12","Type":"ContainerStarted","Data":"e20b9b3ad805b3378f00b56f519bee3a7a9266636a96258528a39123833e60c8"} Feb 20 08:11:10 crc kubenswrapper[4948]: I0220 08:11:10.222850 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tt6t6" podStartSLOduration=2.708221386 podStartE2EDuration="5.222825693s" podCreationTimestamp="2026-02-20 08:11:05 +0000 UTC" firstStartedPulling="2026-02-20 08:11:07.141219431 +0000 UTC m=+316.115714251" lastFinishedPulling="2026-02-20 08:11:09.655823718 +0000 UTC m=+318.630318558" observedRunningTime="2026-02-20 08:11:10.218552008 +0000 UTC m=+319.193046828" watchObservedRunningTime="2026-02-20 08:11:10.222825693 +0000 UTC m=+319.197320553" Feb 20 08:11:10 crc kubenswrapper[4948]: I0220 08:11:10.238563 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dl7nj" podStartSLOduration=2.823887953 podStartE2EDuration="5.238545258s" podCreationTimestamp="2026-02-20 08:11:05 +0000 UTC" firstStartedPulling="2026-02-20 08:11:07.13504133 +0000 UTC m=+316.109536180" lastFinishedPulling="2026-02-20 08:11:09.549698665 +0000 UTC m=+318.524193485" observedRunningTime="2026-02-20 08:11:10.237156434 +0000 UTC m=+319.211651304" watchObservedRunningTime="2026-02-20 08:11:10.238545258 +0000 UTC m=+319.213040078" Feb 20 08:11:11 crc kubenswrapper[4948]: I0220 08:11:11.195933 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x7bcm" event={"ID":"6ec98112-e189-4a27-a6dd-8cb035596810","Type":"ContainerStarted","Data":"fee95fffee0a6e41238f5067917808cd72cf7d7d678b2ec55baeee91c3f20830"} Feb 20 08:11:11 crc kubenswrapper[4948]: I0220 08:11:11.197936 4948 generic.go:334] "Generic (PLEG): container finished" podID="5ee242c9-9783-44cb-a4dc-1ea50a7ae185" containerID="1da87f05a71242abf8441f6ff99031f27ba7460a91f1265f6101e951bc82ac76" exitCode=0 Feb 20 08:11:11 crc kubenswrapper[4948]: I0220 08:11:11.198017 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbtm5" event={"ID":"5ee242c9-9783-44cb-a4dc-1ea50a7ae185","Type":"ContainerDied","Data":"1da87f05a71242abf8441f6ff99031f27ba7460a91f1265f6101e951bc82ac76"} Feb 20 08:11:11 crc kubenswrapper[4948]: I0220 08:11:11.221720 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-x7bcm" podStartSLOduration=2.826067473 podStartE2EDuration="4.221699519s" podCreationTimestamp="2026-02-20 08:11:07 +0000 UTC" firstStartedPulling="2026-02-20 08:11:09.165405741 +0000 UTC m=+318.139900591" lastFinishedPulling="2026-02-20 08:11:10.561037817 +0000 UTC m=+319.535532637" observedRunningTime="2026-02-20 08:11:11.219212198 +0000 UTC m=+320.193707018" watchObservedRunningTime="2026-02-20 08:11:11.221699519 +0000 UTC m=+320.196194349" Feb 20 08:11:11 crc kubenswrapper[4948]: I0220 08:11:11.761039 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-fhfhr"] Feb 20 08:11:11 crc kubenswrapper[4948]: I0220 08:11:11.761746 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" podUID="21db7edd-6db3-4afa-b470-0016a9c5afb9" containerName="controller-manager" containerID="cri-o://aa75719650689fb107484882a72b15a095aba1554265c006218efdf71c2e2e3e" gracePeriod=30 Feb 20 08:11:11 crc kubenswrapper[4948]: I0220 08:11:11.859434 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk"] Feb 20 08:11:11 crc kubenswrapper[4948]: I0220 08:11:11.859725 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" podUID="6c070b67-c2ca-459b-a1b7-813a8833e27e" containerName="route-controller-manager" containerID="cri-o://27c38be8b9e4ccb4d5b9ff6afe53e0b0eb4b257a4accae7cd8daf9b3703d5ce8" gracePeriod=30 Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.172467 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.188482 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21db7edd-6db3-4afa-b470-0016a9c5afb9-serving-cert\") pod \"21db7edd-6db3-4afa-b470-0016a9c5afb9\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.188610 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-proxy-ca-bundles\") pod \"21db7edd-6db3-4afa-b470-0016a9c5afb9\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.188665 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-config\") pod \"21db7edd-6db3-4afa-b470-0016a9c5afb9\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.188774 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-client-ca\") pod \"21db7edd-6db3-4afa-b470-0016a9c5afb9\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.190335 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsbfs\" (UniqueName: \"kubernetes.io/projected/21db7edd-6db3-4afa-b470-0016a9c5afb9-kube-api-access-vsbfs\") pod \"21db7edd-6db3-4afa-b470-0016a9c5afb9\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.190430 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "21db7edd-6db3-4afa-b470-0016a9c5afb9" (UID: "21db7edd-6db3-4afa-b470-0016a9c5afb9"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.190816 4948 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.190985 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-config" (OuterVolumeSpecName: "config") pod "21db7edd-6db3-4afa-b470-0016a9c5afb9" (UID: "21db7edd-6db3-4afa-b470-0016a9c5afb9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.191362 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-client-ca" (OuterVolumeSpecName: "client-ca") pod "21db7edd-6db3-4afa-b470-0016a9c5afb9" (UID: "21db7edd-6db3-4afa-b470-0016a9c5afb9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.197475 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21db7edd-6db3-4afa-b470-0016a9c5afb9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "21db7edd-6db3-4afa-b470-0016a9c5afb9" (UID: "21db7edd-6db3-4afa-b470-0016a9c5afb9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.198571 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21db7edd-6db3-4afa-b470-0016a9c5afb9-kube-api-access-vsbfs" (OuterVolumeSpecName: "kube-api-access-vsbfs") pod "21db7edd-6db3-4afa-b470-0016a9c5afb9" (UID: "21db7edd-6db3-4afa-b470-0016a9c5afb9"). InnerVolumeSpecName "kube-api-access-vsbfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.211479 4948 generic.go:334] "Generic (PLEG): container finished" podID="6c070b67-c2ca-459b-a1b7-813a8833e27e" containerID="27c38be8b9e4ccb4d5b9ff6afe53e0b0eb4b257a4accae7cd8daf9b3703d5ce8" exitCode=0 Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.211562 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" event={"ID":"6c070b67-c2ca-459b-a1b7-813a8833e27e","Type":"ContainerDied","Data":"27c38be8b9e4ccb4d5b9ff6afe53e0b0eb4b257a4accae7cd8daf9b3703d5ce8"} Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.213445 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbtm5" event={"ID":"5ee242c9-9783-44cb-a4dc-1ea50a7ae185","Type":"ContainerStarted","Data":"3b29859986cd536efddb95cc16c282cbd1fab4057f4305e4757cce50ace32d1a"} Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.226841 4948 generic.go:334] "Generic (PLEG): container finished" podID="21db7edd-6db3-4afa-b470-0016a9c5afb9" containerID="aa75719650689fb107484882a72b15a095aba1554265c006218efdf71c2e2e3e" exitCode=0 Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.227258 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" event={"ID":"21db7edd-6db3-4afa-b470-0016a9c5afb9","Type":"ContainerDied","Data":"aa75719650689fb107484882a72b15a095aba1554265c006218efdf71c2e2e3e"} Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.227334 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" event={"ID":"21db7edd-6db3-4afa-b470-0016a9c5afb9","Type":"ContainerDied","Data":"84341e8559dd79e093cb4de6142b9a22ac6a50040f68f50623fb6ec98aa5e33a"} Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.227360 4948 scope.go:117] "RemoveContainer" containerID="aa75719650689fb107484882a72b15a095aba1554265c006218efdf71c2e2e3e" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.227210 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.250090 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jbtm5" podStartSLOduration=1.844780842 podStartE2EDuration="4.250070229s" podCreationTimestamp="2026-02-20 08:11:08 +0000 UTC" firstStartedPulling="2026-02-20 08:11:09.175053347 +0000 UTC m=+318.149548177" lastFinishedPulling="2026-02-20 08:11:11.580342744 +0000 UTC m=+320.554837564" observedRunningTime="2026-02-20 08:11:12.247366012 +0000 UTC m=+321.221860832" watchObservedRunningTime="2026-02-20 08:11:12.250070229 +0000 UTC m=+321.224565049" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.254941 4948 scope.go:117] "RemoveContainer" containerID="aa75719650689fb107484882a72b15a095aba1554265c006218efdf71c2e2e3e" Feb 20 08:11:12 crc kubenswrapper[4948]: E0220 08:11:12.255722 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa75719650689fb107484882a72b15a095aba1554265c006218efdf71c2e2e3e\": container with ID starting with aa75719650689fb107484882a72b15a095aba1554265c006218efdf71c2e2e3e not found: ID does not exist" containerID="aa75719650689fb107484882a72b15a095aba1554265c006218efdf71c2e2e3e" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.255757 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa75719650689fb107484882a72b15a095aba1554265c006218efdf71c2e2e3e"} err="failed to get container status \"aa75719650689fb107484882a72b15a095aba1554265c006218efdf71c2e2e3e\": rpc error: code = NotFound desc = could not find container \"aa75719650689fb107484882a72b15a095aba1554265c006218efdf71c2e2e3e\": container with ID starting with aa75719650689fb107484882a72b15a095aba1554265c006218efdf71c2e2e3e not found: ID does not exist" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.256757 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.285659 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-fhfhr"] Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.292027 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.292053 4948 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-client-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.292066 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsbfs\" (UniqueName: \"kubernetes.io/projected/21db7edd-6db3-4afa-b470-0016a9c5afb9-kube-api-access-vsbfs\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.292076 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21db7edd-6db3-4afa-b470-0016a9c5afb9-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.299051 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-fhfhr"] Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.392419 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c070b67-c2ca-459b-a1b7-813a8833e27e-config\") pod \"6c070b67-c2ca-459b-a1b7-813a8833e27e\" (UID: \"6c070b67-c2ca-459b-a1b7-813a8833e27e\") " Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.392476 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c070b67-c2ca-459b-a1b7-813a8833e27e-serving-cert\") pod \"6c070b67-c2ca-459b-a1b7-813a8833e27e\" (UID: \"6c070b67-c2ca-459b-a1b7-813a8833e27e\") " Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.392516 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6c070b67-c2ca-459b-a1b7-813a8833e27e-client-ca\") pod \"6c070b67-c2ca-459b-a1b7-813a8833e27e\" (UID: \"6c070b67-c2ca-459b-a1b7-813a8833e27e\") " Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.392594 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25nlb\" (UniqueName: \"kubernetes.io/projected/6c070b67-c2ca-459b-a1b7-813a8833e27e-kube-api-access-25nlb\") pod \"6c070b67-c2ca-459b-a1b7-813a8833e27e\" (UID: \"6c070b67-c2ca-459b-a1b7-813a8833e27e\") " Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.393995 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c070b67-c2ca-459b-a1b7-813a8833e27e-client-ca" (OuterVolumeSpecName: "client-ca") pod "6c070b67-c2ca-459b-a1b7-813a8833e27e" (UID: "6c070b67-c2ca-459b-a1b7-813a8833e27e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.394185 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c070b67-c2ca-459b-a1b7-813a8833e27e-config" (OuterVolumeSpecName: "config") pod "6c070b67-c2ca-459b-a1b7-813a8833e27e" (UID: "6c070b67-c2ca-459b-a1b7-813a8833e27e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.397815 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c070b67-c2ca-459b-a1b7-813a8833e27e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6c070b67-c2ca-459b-a1b7-813a8833e27e" (UID: "6c070b67-c2ca-459b-a1b7-813a8833e27e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.398197 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c070b67-c2ca-459b-a1b7-813a8833e27e-kube-api-access-25nlb" (OuterVolumeSpecName: "kube-api-access-25nlb") pod "6c070b67-c2ca-459b-a1b7-813a8833e27e" (UID: "6c070b67-c2ca-459b-a1b7-813a8833e27e"). InnerVolumeSpecName "kube-api-access-25nlb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.493706 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25nlb\" (UniqueName: \"kubernetes.io/projected/6c070b67-c2ca-459b-a1b7-813a8833e27e-kube-api-access-25nlb\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.493744 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c070b67-c2ca-459b-a1b7-813a8833e27e-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.493754 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c070b67-c2ca-459b-a1b7-813a8833e27e-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.493762 4948 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6c070b67-c2ca-459b-a1b7-813a8833e27e-client-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.237514 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.237526 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" event={"ID":"6c070b67-c2ca-459b-a1b7-813a8833e27e","Type":"ContainerDied","Data":"f24acc5a18bb988f4e5c8c67e817d3ea2636025712e6b3477a09950d4595e36b"} Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.237622 4948 scope.go:117] "RemoveContainer" containerID="27c38be8b9e4ccb4d5b9ff6afe53e0b0eb4b257a4accae7cd8daf9b3703d5ce8" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.296504 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk"] Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.302468 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk"] Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.346352 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5dfff44d5b-x9695"] Feb 20 08:11:13 crc kubenswrapper[4948]: E0220 08:11:13.346665 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21db7edd-6db3-4afa-b470-0016a9c5afb9" containerName="controller-manager" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.346687 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="21db7edd-6db3-4afa-b470-0016a9c5afb9" containerName="controller-manager" Feb 20 08:11:13 crc kubenswrapper[4948]: E0220 08:11:13.346729 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c070b67-c2ca-459b-a1b7-813a8833e27e" containerName="route-controller-manager" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.346737 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c070b67-c2ca-459b-a1b7-813a8833e27e" containerName="route-controller-manager" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.346857 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c070b67-c2ca-459b-a1b7-813a8833e27e" containerName="route-controller-manager" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.346881 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="21db7edd-6db3-4afa-b470-0016a9c5afb9" containerName="controller-manager" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.347401 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.350250 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.351077 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.351653 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.354800 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.355031 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.356852 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.360434 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4"] Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.361261 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.363926 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.364002 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.364469 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.364998 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.365369 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.365833 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.368324 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.369935 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5dfff44d5b-x9695"] Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.376523 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4"] Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.506921 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-config\") pod \"route-controller-manager-c99b4965f-dm2z4\" (UID: \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.507033 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fb83d65b-730e-4024-a9a8-27b2a18b305d-proxy-ca-bundles\") pod \"controller-manager-5dfff44d5b-x9695\" (UID: \"fb83d65b-730e-4024-a9a8-27b2a18b305d\") " pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.507113 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fb83d65b-730e-4024-a9a8-27b2a18b305d-client-ca\") pod \"controller-manager-5dfff44d5b-x9695\" (UID: \"fb83d65b-730e-4024-a9a8-27b2a18b305d\") " pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.507229 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb83d65b-730e-4024-a9a8-27b2a18b305d-serving-cert\") pod \"controller-manager-5dfff44d5b-x9695\" (UID: \"fb83d65b-730e-4024-a9a8-27b2a18b305d\") " pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.507273 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-serving-cert\") pod \"route-controller-manager-c99b4965f-dm2z4\" (UID: \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.507327 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb83d65b-730e-4024-a9a8-27b2a18b305d-config\") pod \"controller-manager-5dfff44d5b-x9695\" (UID: \"fb83d65b-730e-4024-a9a8-27b2a18b305d\") " pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.507442 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljpsl\" (UniqueName: \"kubernetes.io/projected/fb83d65b-730e-4024-a9a8-27b2a18b305d-kube-api-access-ljpsl\") pod \"controller-manager-5dfff44d5b-x9695\" (UID: \"fb83d65b-730e-4024-a9a8-27b2a18b305d\") " pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.507480 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jng5g\" (UniqueName: \"kubernetes.io/projected/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-kube-api-access-jng5g\") pod \"route-controller-manager-c99b4965f-dm2z4\" (UID: \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.507529 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-client-ca\") pod \"route-controller-manager-c99b4965f-dm2z4\" (UID: \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.608415 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-config\") pod \"route-controller-manager-c99b4965f-dm2z4\" (UID: \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.608508 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fb83d65b-730e-4024-a9a8-27b2a18b305d-proxy-ca-bundles\") pod \"controller-manager-5dfff44d5b-x9695\" (UID: \"fb83d65b-730e-4024-a9a8-27b2a18b305d\") " pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.608566 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fb83d65b-730e-4024-a9a8-27b2a18b305d-client-ca\") pod \"controller-manager-5dfff44d5b-x9695\" (UID: \"fb83d65b-730e-4024-a9a8-27b2a18b305d\") " pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.608621 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb83d65b-730e-4024-a9a8-27b2a18b305d-serving-cert\") pod \"controller-manager-5dfff44d5b-x9695\" (UID: \"fb83d65b-730e-4024-a9a8-27b2a18b305d\") " pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.608660 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-serving-cert\") pod \"route-controller-manager-c99b4965f-dm2z4\" (UID: \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.608701 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb83d65b-730e-4024-a9a8-27b2a18b305d-config\") pod \"controller-manager-5dfff44d5b-x9695\" (UID: \"fb83d65b-730e-4024-a9a8-27b2a18b305d\") " pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.608758 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljpsl\" (UniqueName: \"kubernetes.io/projected/fb83d65b-730e-4024-a9a8-27b2a18b305d-kube-api-access-ljpsl\") pod \"controller-manager-5dfff44d5b-x9695\" (UID: \"fb83d65b-730e-4024-a9a8-27b2a18b305d\") " pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.608792 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jng5g\" (UniqueName: \"kubernetes.io/projected/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-kube-api-access-jng5g\") pod \"route-controller-manager-c99b4965f-dm2z4\" (UID: \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.608862 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-client-ca\") pod \"route-controller-manager-c99b4965f-dm2z4\" (UID: \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.609987 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-config\") pod \"route-controller-manager-c99b4965f-dm2z4\" (UID: \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.610366 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-client-ca\") pod \"route-controller-manager-c99b4965f-dm2z4\" (UID: \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.610478 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fb83d65b-730e-4024-a9a8-27b2a18b305d-client-ca\") pod \"controller-manager-5dfff44d5b-x9695\" (UID: \"fb83d65b-730e-4024-a9a8-27b2a18b305d\") " pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.610602 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb83d65b-730e-4024-a9a8-27b2a18b305d-config\") pod \"controller-manager-5dfff44d5b-x9695\" (UID: \"fb83d65b-730e-4024-a9a8-27b2a18b305d\") " pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.611286 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fb83d65b-730e-4024-a9a8-27b2a18b305d-proxy-ca-bundles\") pod \"controller-manager-5dfff44d5b-x9695\" (UID: \"fb83d65b-730e-4024-a9a8-27b2a18b305d\") " pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.614502 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-serving-cert\") pod \"route-controller-manager-c99b4965f-dm2z4\" (UID: \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.614692 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb83d65b-730e-4024-a9a8-27b2a18b305d-serving-cert\") pod \"controller-manager-5dfff44d5b-x9695\" (UID: \"fb83d65b-730e-4024-a9a8-27b2a18b305d\") " pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.631706 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jng5g\" (UniqueName: \"kubernetes.io/projected/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-kube-api-access-jng5g\") pod \"route-controller-manager-c99b4965f-dm2z4\" (UID: \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.642514 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljpsl\" (UniqueName: \"kubernetes.io/projected/fb83d65b-730e-4024-a9a8-27b2a18b305d-kube-api-access-ljpsl\") pod \"controller-manager-5dfff44d5b-x9695\" (UID: \"fb83d65b-730e-4024-a9a8-27b2a18b305d\") " pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.683440 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.700084 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.729416 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21db7edd-6db3-4afa-b470-0016a9c5afb9" path="/var/lib/kubelet/pods/21db7edd-6db3-4afa-b470-0016a9c5afb9/volumes" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.730400 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c070b67-c2ca-459b-a1b7-813a8833e27e" path="/var/lib/kubelet/pods/6c070b67-c2ca-459b-a1b7-813a8833e27e/volumes" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.952777 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4"] Feb 20 08:11:13 crc kubenswrapper[4948]: W0220 08:11:13.962227 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2fba2cd6_ab12_40bb_90e7_17681ddc5bcf.slice/crio-b7b0531bcc9a09086ceb7e7ef51d7fe67486fce64c4091ce69022bf2050f95fe WatchSource:0}: Error finding container b7b0531bcc9a09086ceb7e7ef51d7fe67486fce64c4091ce69022bf2050f95fe: Status 404 returned error can't find the container with id b7b0531bcc9a09086ceb7e7ef51d7fe67486fce64c4091ce69022bf2050f95fe Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.990726 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5dfff44d5b-x9695"] Feb 20 08:11:13 crc kubenswrapper[4948]: W0220 08:11:13.998101 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfb83d65b_730e_4024_a9a8_27b2a18b305d.slice/crio-3390c540d19ba441062645bacaa2af7d0d4366dcacd7de98b84bc73030b8affd WatchSource:0}: Error finding container 3390c540d19ba441062645bacaa2af7d0d4366dcacd7de98b84bc73030b8affd: Status 404 returned error can't find the container with id 3390c540d19ba441062645bacaa2af7d0d4366dcacd7de98b84bc73030b8affd Feb 20 08:11:14 crc kubenswrapper[4948]: I0220 08:11:14.245944 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" event={"ID":"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf","Type":"ContainerStarted","Data":"e01ead6dca19e54015aea007496c57209ea14c70dd72f890df0bf653660cc282"} Feb 20 08:11:14 crc kubenswrapper[4948]: I0220 08:11:14.246016 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" event={"ID":"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf","Type":"ContainerStarted","Data":"b7b0531bcc9a09086ceb7e7ef51d7fe67486fce64c4091ce69022bf2050f95fe"} Feb 20 08:11:14 crc kubenswrapper[4948]: I0220 08:11:14.246490 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:14 crc kubenswrapper[4948]: I0220 08:11:14.247617 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" event={"ID":"fb83d65b-730e-4024-a9a8-27b2a18b305d","Type":"ContainerStarted","Data":"ccba885eaf6bc7f27fba8d3843893c2538dc7e82d238fe2e090637adf95f2666"} Feb 20 08:11:14 crc kubenswrapper[4948]: I0220 08:11:14.247648 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" event={"ID":"fb83d65b-730e-4024-a9a8-27b2a18b305d","Type":"ContainerStarted","Data":"3390c540d19ba441062645bacaa2af7d0d4366dcacd7de98b84bc73030b8affd"} Feb 20 08:11:14 crc kubenswrapper[4948]: I0220 08:11:14.248118 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:14 crc kubenswrapper[4948]: I0220 08:11:14.252830 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:14 crc kubenswrapper[4948]: I0220 08:11:14.262531 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" podStartSLOduration=3.26251091 podStartE2EDuration="3.26251091s" podCreationTimestamp="2026-02-20 08:11:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:11:14.2604499 +0000 UTC m=+323.234944760" watchObservedRunningTime="2026-02-20 08:11:14.26251091 +0000 UTC m=+323.237005730" Feb 20 08:11:14 crc kubenswrapper[4948]: I0220 08:11:14.492838 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:14 crc kubenswrapper[4948]: I0220 08:11:14.522515 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" podStartSLOduration=3.522500976 podStartE2EDuration="3.522500976s" podCreationTimestamp="2026-02-20 08:11:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:11:14.281374703 +0000 UTC m=+323.255869543" watchObservedRunningTime="2026-02-20 08:11:14.522500976 +0000 UTC m=+323.496995796" Feb 20 08:11:15 crc kubenswrapper[4948]: I0220 08:11:15.382034 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4"] Feb 20 08:11:15 crc kubenswrapper[4948]: I0220 08:11:15.845204 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dl7nj" Feb 20 08:11:15 crc kubenswrapper[4948]: I0220 08:11:15.845284 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dl7nj" Feb 20 08:11:15 crc kubenswrapper[4948]: I0220 08:11:15.906289 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dl7nj" Feb 20 08:11:15 crc kubenswrapper[4948]: I0220 08:11:15.991173 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tt6t6" Feb 20 08:11:15 crc kubenswrapper[4948]: I0220 08:11:15.991532 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tt6t6" Feb 20 08:11:16 crc kubenswrapper[4948]: I0220 08:11:16.039740 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tt6t6" Feb 20 08:11:16 crc kubenswrapper[4948]: I0220 08:11:16.308209 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tt6t6" Feb 20 08:11:16 crc kubenswrapper[4948]: I0220 08:11:16.319307 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dl7nj" Feb 20 08:11:17 crc kubenswrapper[4948]: I0220 08:11:17.268146 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" podUID="2fba2cd6-ab12-40bb-90e7-17681ddc5bcf" containerName="route-controller-manager" containerID="cri-o://e01ead6dca19e54015aea007496c57209ea14c70dd72f890df0bf653660cc282" gracePeriod=30 Feb 20 08:11:18 crc kubenswrapper[4948]: I0220 08:11:18.216966 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-x7bcm" Feb 20 08:11:18 crc kubenswrapper[4948]: I0220 08:11:18.217041 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-x7bcm" Feb 20 08:11:18 crc kubenswrapper[4948]: I0220 08:11:18.262687 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-x7bcm" Feb 20 08:11:18 crc kubenswrapper[4948]: I0220 08:11:18.344800 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-x7bcm" Feb 20 08:11:18 crc kubenswrapper[4948]: I0220 08:11:18.467626 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jbtm5" Feb 20 08:11:18 crc kubenswrapper[4948]: I0220 08:11:18.467682 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jbtm5" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.281397 4948 generic.go:334] "Generic (PLEG): container finished" podID="2fba2cd6-ab12-40bb-90e7-17681ddc5bcf" containerID="e01ead6dca19e54015aea007496c57209ea14c70dd72f890df0bf653660cc282" exitCode=0 Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.281540 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" event={"ID":"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf","Type":"ContainerDied","Data":"e01ead6dca19e54015aea007496c57209ea14c70dd72f890df0bf653660cc282"} Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.501081 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jbtm5" podUID="5ee242c9-9783-44cb-a4dc-1ea50a7ae185" containerName="registry-server" probeResult="failure" output=< Feb 20 08:11:19 crc kubenswrapper[4948]: timeout: failed to connect service ":50051" within 1s Feb 20 08:11:19 crc kubenswrapper[4948]: > Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.689054 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.717652 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m"] Feb 20 08:11:19 crc kubenswrapper[4948]: E0220 08:11:19.717998 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fba2cd6-ab12-40bb-90e7-17681ddc5bcf" containerName="route-controller-manager" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.718020 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fba2cd6-ab12-40bb-90e7-17681ddc5bcf" containerName="route-controller-manager" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.718257 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fba2cd6-ab12-40bb-90e7-17681ddc5bcf" containerName="route-controller-manager" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.718839 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.736849 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m"] Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.859846 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jng5g\" (UniqueName: \"kubernetes.io/projected/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-kube-api-access-jng5g\") pod \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\" (UID: \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\") " Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.860340 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-client-ca\") pod \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\" (UID: \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\") " Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.860512 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-serving-cert\") pod \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\" (UID: \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\") " Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.860680 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-config\") pod \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\" (UID: \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\") " Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.860930 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8p8l\" (UniqueName: \"kubernetes.io/projected/7d1617e1-d79c-4197-ac74-edc70e60870d-kube-api-access-z8p8l\") pod \"route-controller-manager-7565cf79bc-8x44m\" (UID: \"7d1617e1-d79c-4197-ac74-edc70e60870d\") " pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.861061 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d1617e1-d79c-4197-ac74-edc70e60870d-serving-cert\") pod \"route-controller-manager-7565cf79bc-8x44m\" (UID: \"7d1617e1-d79c-4197-ac74-edc70e60870d\") " pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.861217 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7d1617e1-d79c-4197-ac74-edc70e60870d-client-ca\") pod \"route-controller-manager-7565cf79bc-8x44m\" (UID: \"7d1617e1-d79c-4197-ac74-edc70e60870d\") " pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.861226 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-client-ca" (OuterVolumeSpecName: "client-ca") pod "2fba2cd6-ab12-40bb-90e7-17681ddc5bcf" (UID: "2fba2cd6-ab12-40bb-90e7-17681ddc5bcf"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.861365 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-config" (OuterVolumeSpecName: "config") pod "2fba2cd6-ab12-40bb-90e7-17681ddc5bcf" (UID: "2fba2cd6-ab12-40bb-90e7-17681ddc5bcf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.861684 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d1617e1-d79c-4197-ac74-edc70e60870d-config\") pod \"route-controller-manager-7565cf79bc-8x44m\" (UID: \"7d1617e1-d79c-4197-ac74-edc70e60870d\") " pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.861870 4948 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-client-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.861892 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.869246 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-kube-api-access-jng5g" (OuterVolumeSpecName: "kube-api-access-jng5g") pod "2fba2cd6-ab12-40bb-90e7-17681ddc5bcf" (UID: "2fba2cd6-ab12-40bb-90e7-17681ddc5bcf"). InnerVolumeSpecName "kube-api-access-jng5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.869508 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "2fba2cd6-ab12-40bb-90e7-17681ddc5bcf" (UID: "2fba2cd6-ab12-40bb-90e7-17681ddc5bcf"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.963041 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8p8l\" (UniqueName: \"kubernetes.io/projected/7d1617e1-d79c-4197-ac74-edc70e60870d-kube-api-access-z8p8l\") pod \"route-controller-manager-7565cf79bc-8x44m\" (UID: \"7d1617e1-d79c-4197-ac74-edc70e60870d\") " pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.963092 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d1617e1-d79c-4197-ac74-edc70e60870d-serving-cert\") pod \"route-controller-manager-7565cf79bc-8x44m\" (UID: \"7d1617e1-d79c-4197-ac74-edc70e60870d\") " pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.963131 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7d1617e1-d79c-4197-ac74-edc70e60870d-client-ca\") pod \"route-controller-manager-7565cf79bc-8x44m\" (UID: \"7d1617e1-d79c-4197-ac74-edc70e60870d\") " pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.963151 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d1617e1-d79c-4197-ac74-edc70e60870d-config\") pod \"route-controller-manager-7565cf79bc-8x44m\" (UID: \"7d1617e1-d79c-4197-ac74-edc70e60870d\") " pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.963186 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jng5g\" (UniqueName: \"kubernetes.io/projected/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-kube-api-access-jng5g\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.963201 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.964324 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7d1617e1-d79c-4197-ac74-edc70e60870d-client-ca\") pod \"route-controller-manager-7565cf79bc-8x44m\" (UID: \"7d1617e1-d79c-4197-ac74-edc70e60870d\") " pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.964524 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d1617e1-d79c-4197-ac74-edc70e60870d-config\") pod \"route-controller-manager-7565cf79bc-8x44m\" (UID: \"7d1617e1-d79c-4197-ac74-edc70e60870d\") " pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.966953 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d1617e1-d79c-4197-ac74-edc70e60870d-serving-cert\") pod \"route-controller-manager-7565cf79bc-8x44m\" (UID: \"7d1617e1-d79c-4197-ac74-edc70e60870d\") " pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.982018 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8p8l\" (UniqueName: \"kubernetes.io/projected/7d1617e1-d79c-4197-ac74-edc70e60870d-kube-api-access-z8p8l\") pod \"route-controller-manager-7565cf79bc-8x44m\" (UID: \"7d1617e1-d79c-4197-ac74-edc70e60870d\") " pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:20 crc kubenswrapper[4948]: I0220 08:11:20.054440 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:20 crc kubenswrapper[4948]: I0220 08:11:20.288638 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:20 crc kubenswrapper[4948]: I0220 08:11:20.291381 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" event={"ID":"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf","Type":"ContainerDied","Data":"b7b0531bcc9a09086ceb7e7ef51d7fe67486fce64c4091ce69022bf2050f95fe"} Feb 20 08:11:20 crc kubenswrapper[4948]: I0220 08:11:20.291484 4948 scope.go:117] "RemoveContainer" containerID="e01ead6dca19e54015aea007496c57209ea14c70dd72f890df0bf653660cc282" Feb 20 08:11:20 crc kubenswrapper[4948]: I0220 08:11:20.317735 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4"] Feb 20 08:11:20 crc kubenswrapper[4948]: I0220 08:11:20.321208 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4"] Feb 20 08:11:20 crc kubenswrapper[4948]: I0220 08:11:20.472157 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m"] Feb 20 08:11:21 crc kubenswrapper[4948]: I0220 08:11:21.295775 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" event={"ID":"7d1617e1-d79c-4197-ac74-edc70e60870d","Type":"ContainerStarted","Data":"64ad04f585e964b7677390442b789797ba61c7d5f4d2e507fba783bd201bcc36"} Feb 20 08:11:21 crc kubenswrapper[4948]: I0220 08:11:21.296166 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" event={"ID":"7d1617e1-d79c-4197-ac74-edc70e60870d","Type":"ContainerStarted","Data":"946f59be51177e033bd3a18511a444f58bf8d8e903a9bda2540b30b85ad59a9f"} Feb 20 08:11:21 crc kubenswrapper[4948]: I0220 08:11:21.296470 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:21 crc kubenswrapper[4948]: I0220 08:11:21.304789 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:21 crc kubenswrapper[4948]: I0220 08:11:21.321774 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" podStartSLOduration=6.321758095 podStartE2EDuration="6.321758095s" podCreationTimestamp="2026-02-20 08:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:11:21.321257632 +0000 UTC m=+330.295752462" watchObservedRunningTime="2026-02-20 08:11:21.321758095 +0000 UTC m=+330.296252935" Feb 20 08:11:21 crc kubenswrapper[4948]: I0220 08:11:21.727869 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fba2cd6-ab12-40bb-90e7-17681ddc5bcf" path="/var/lib/kubelet/pods/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf/volumes" Feb 20 08:11:28 crc kubenswrapper[4948]: I0220 08:11:28.537164 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jbtm5" Feb 20 08:11:28 crc kubenswrapper[4948]: I0220 08:11:28.599875 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jbtm5" Feb 20 08:11:38 crc kubenswrapper[4948]: I0220 08:11:38.025441 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:11:38 crc kubenswrapper[4948]: I0220 08:11:38.026733 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:11:46 crc kubenswrapper[4948]: I0220 08:11:46.879894 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-hhhfd"] Feb 20 08:11:46 crc kubenswrapper[4948]: I0220 08:11:46.882122 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:46 crc kubenswrapper[4948]: I0220 08:11:46.905318 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-hhhfd"] Feb 20 08:11:46 crc kubenswrapper[4948]: I0220 08:11:46.987120 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d804474b-3915-46f1-9fc7-366ccfb6602b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:46 crc kubenswrapper[4948]: I0220 08:11:46.987175 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d804474b-3915-46f1-9fc7-366ccfb6602b-registry-tls\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:46 crc kubenswrapper[4948]: I0220 08:11:46.987212 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:46 crc kubenswrapper[4948]: I0220 08:11:46.987242 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d804474b-3915-46f1-9fc7-366ccfb6602b-trusted-ca\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:46 crc kubenswrapper[4948]: I0220 08:11:46.987340 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d804474b-3915-46f1-9fc7-366ccfb6602b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:46 crc kubenswrapper[4948]: I0220 08:11:46.987364 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d804474b-3915-46f1-9fc7-366ccfb6602b-bound-sa-token\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:46 crc kubenswrapper[4948]: I0220 08:11:46.987394 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d804474b-3915-46f1-9fc7-366ccfb6602b-registry-certificates\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:46 crc kubenswrapper[4948]: I0220 08:11:46.987431 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbnlk\" (UniqueName: \"kubernetes.io/projected/d804474b-3915-46f1-9fc7-366ccfb6602b-kube-api-access-hbnlk\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:47 crc kubenswrapper[4948]: I0220 08:11:47.011450 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:47 crc kubenswrapper[4948]: I0220 08:11:47.089154 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d804474b-3915-46f1-9fc7-366ccfb6602b-registry-certificates\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:47 crc kubenswrapper[4948]: I0220 08:11:47.089225 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbnlk\" (UniqueName: \"kubernetes.io/projected/d804474b-3915-46f1-9fc7-366ccfb6602b-kube-api-access-hbnlk\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:47 crc kubenswrapper[4948]: I0220 08:11:47.089291 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d804474b-3915-46f1-9fc7-366ccfb6602b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:47 crc kubenswrapper[4948]: I0220 08:11:47.089320 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d804474b-3915-46f1-9fc7-366ccfb6602b-registry-tls\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:47 crc kubenswrapper[4948]: I0220 08:11:47.089349 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d804474b-3915-46f1-9fc7-366ccfb6602b-trusted-ca\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:47 crc kubenswrapper[4948]: I0220 08:11:47.089393 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d804474b-3915-46f1-9fc7-366ccfb6602b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:47 crc kubenswrapper[4948]: I0220 08:11:47.089414 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d804474b-3915-46f1-9fc7-366ccfb6602b-bound-sa-token\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:47 crc kubenswrapper[4948]: I0220 08:11:47.091056 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d804474b-3915-46f1-9fc7-366ccfb6602b-registry-certificates\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:47 crc kubenswrapper[4948]: I0220 08:11:47.095353 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d804474b-3915-46f1-9fc7-366ccfb6602b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:47 crc kubenswrapper[4948]: I0220 08:11:47.112765 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d804474b-3915-46f1-9fc7-366ccfb6602b-trusted-ca\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:47 crc kubenswrapper[4948]: I0220 08:11:47.117347 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d804474b-3915-46f1-9fc7-366ccfb6602b-registry-tls\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:47 crc kubenswrapper[4948]: I0220 08:11:47.119255 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbnlk\" (UniqueName: \"kubernetes.io/projected/d804474b-3915-46f1-9fc7-366ccfb6602b-kube-api-access-hbnlk\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:47 crc kubenswrapper[4948]: I0220 08:11:47.121572 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d804474b-3915-46f1-9fc7-366ccfb6602b-bound-sa-token\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:47 crc kubenswrapper[4948]: I0220 08:11:47.121631 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d804474b-3915-46f1-9fc7-366ccfb6602b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:47 crc kubenswrapper[4948]: I0220 08:11:47.204514 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:47 crc kubenswrapper[4948]: I0220 08:11:47.660733 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-hhhfd"] Feb 20 08:11:48 crc kubenswrapper[4948]: I0220 08:11:48.491171 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" event={"ID":"d804474b-3915-46f1-9fc7-366ccfb6602b","Type":"ContainerStarted","Data":"96395697398b7e725b35e6e3db0a9fecbe5007e3babb791ce6d7d562c98144b0"} Feb 20 08:11:48 crc kubenswrapper[4948]: I0220 08:11:48.491655 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:48 crc kubenswrapper[4948]: I0220 08:11:48.491672 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" event={"ID":"d804474b-3915-46f1-9fc7-366ccfb6602b","Type":"ContainerStarted","Data":"bc2abfbb23d88c80d5df543d5cf4ecd494a9e31b5c733d2e7c759003511f1ec0"} Feb 20 08:11:48 crc kubenswrapper[4948]: I0220 08:11:48.520359 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" podStartSLOduration=2.520324075 podStartE2EDuration="2.520324075s" podCreationTimestamp="2026-02-20 08:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:11:48.520023727 +0000 UTC m=+357.494518547" watchObservedRunningTime="2026-02-20 08:11:48.520324075 +0000 UTC m=+357.494818935" Feb 20 08:11:51 crc kubenswrapper[4948]: I0220 08:11:51.756063 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m"] Feb 20 08:11:51 crc kubenswrapper[4948]: I0220 08:11:51.757745 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" podUID="7d1617e1-d79c-4197-ac74-edc70e60870d" containerName="route-controller-manager" containerID="cri-o://64ad04f585e964b7677390442b789797ba61c7d5f4d2e507fba783bd201bcc36" gracePeriod=30 Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.212514 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.279527 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d1617e1-d79c-4197-ac74-edc70e60870d-serving-cert\") pod \"7d1617e1-d79c-4197-ac74-edc70e60870d\" (UID: \"7d1617e1-d79c-4197-ac74-edc70e60870d\") " Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.279610 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d1617e1-d79c-4197-ac74-edc70e60870d-config\") pod \"7d1617e1-d79c-4197-ac74-edc70e60870d\" (UID: \"7d1617e1-d79c-4197-ac74-edc70e60870d\") " Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.279669 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8p8l\" (UniqueName: \"kubernetes.io/projected/7d1617e1-d79c-4197-ac74-edc70e60870d-kube-api-access-z8p8l\") pod \"7d1617e1-d79c-4197-ac74-edc70e60870d\" (UID: \"7d1617e1-d79c-4197-ac74-edc70e60870d\") " Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.279785 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7d1617e1-d79c-4197-ac74-edc70e60870d-client-ca\") pod \"7d1617e1-d79c-4197-ac74-edc70e60870d\" (UID: \"7d1617e1-d79c-4197-ac74-edc70e60870d\") " Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.280256 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d1617e1-d79c-4197-ac74-edc70e60870d-config" (OuterVolumeSpecName: "config") pod "7d1617e1-d79c-4197-ac74-edc70e60870d" (UID: "7d1617e1-d79c-4197-ac74-edc70e60870d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.280398 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d1617e1-d79c-4197-ac74-edc70e60870d-client-ca" (OuterVolumeSpecName: "client-ca") pod "7d1617e1-d79c-4197-ac74-edc70e60870d" (UID: "7d1617e1-d79c-4197-ac74-edc70e60870d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.285876 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d1617e1-d79c-4197-ac74-edc70e60870d-kube-api-access-z8p8l" (OuterVolumeSpecName: "kube-api-access-z8p8l") pod "7d1617e1-d79c-4197-ac74-edc70e60870d" (UID: "7d1617e1-d79c-4197-ac74-edc70e60870d"). InnerVolumeSpecName "kube-api-access-z8p8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.288921 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d1617e1-d79c-4197-ac74-edc70e60870d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7d1617e1-d79c-4197-ac74-edc70e60870d" (UID: "7d1617e1-d79c-4197-ac74-edc70e60870d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.381846 4948 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7d1617e1-d79c-4197-ac74-edc70e60870d-client-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.381887 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d1617e1-d79c-4197-ac74-edc70e60870d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.381902 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d1617e1-d79c-4197-ac74-edc70e60870d-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.381914 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8p8l\" (UniqueName: \"kubernetes.io/projected/7d1617e1-d79c-4197-ac74-edc70e60870d-kube-api-access-z8p8l\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.517438 4948 generic.go:334] "Generic (PLEG): container finished" podID="7d1617e1-d79c-4197-ac74-edc70e60870d" containerID="64ad04f585e964b7677390442b789797ba61c7d5f4d2e507fba783bd201bcc36" exitCode=0 Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.517488 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" event={"ID":"7d1617e1-d79c-4197-ac74-edc70e60870d","Type":"ContainerDied","Data":"64ad04f585e964b7677390442b789797ba61c7d5f4d2e507fba783bd201bcc36"} Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.517515 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" event={"ID":"7d1617e1-d79c-4197-ac74-edc70e60870d","Type":"ContainerDied","Data":"946f59be51177e033bd3a18511a444f58bf8d8e903a9bda2540b30b85ad59a9f"} Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.517542 4948 scope.go:117] "RemoveContainer" containerID="64ad04f585e964b7677390442b789797ba61c7d5f4d2e507fba783bd201bcc36" Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.517544 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.535890 4948 scope.go:117] "RemoveContainer" containerID="64ad04f585e964b7677390442b789797ba61c7d5f4d2e507fba783bd201bcc36" Feb 20 08:11:52 crc kubenswrapper[4948]: E0220 08:11:52.536580 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64ad04f585e964b7677390442b789797ba61c7d5f4d2e507fba783bd201bcc36\": container with ID starting with 64ad04f585e964b7677390442b789797ba61c7d5f4d2e507fba783bd201bcc36 not found: ID does not exist" containerID="64ad04f585e964b7677390442b789797ba61c7d5f4d2e507fba783bd201bcc36" Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.536645 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64ad04f585e964b7677390442b789797ba61c7d5f4d2e507fba783bd201bcc36"} err="failed to get container status \"64ad04f585e964b7677390442b789797ba61c7d5f4d2e507fba783bd201bcc36\": rpc error: code = NotFound desc = could not find container \"64ad04f585e964b7677390442b789797ba61c7d5f4d2e507fba783bd201bcc36\": container with ID starting with 64ad04f585e964b7677390442b789797ba61c7d5f4d2e507fba783bd201bcc36 not found: ID does not exist" Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.551790 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m"] Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.559939 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m"] Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.368316 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9"] Feb 20 08:11:53 crc kubenswrapper[4948]: E0220 08:11:53.368878 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d1617e1-d79c-4197-ac74-edc70e60870d" containerName="route-controller-manager" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.368896 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d1617e1-d79c-4197-ac74-edc70e60870d" containerName="route-controller-manager" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.369028 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d1617e1-d79c-4197-ac74-edc70e60870d" containerName="route-controller-manager" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.369479 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.371958 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.372834 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.373208 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.373411 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.373576 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.374023 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.386469 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9"] Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.496146 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05df7d38-ed40-42b4-9761-0ff40846f561-config\") pod \"route-controller-manager-c99b4965f-q7mv9\" (UID: \"05df7d38-ed40-42b4-9761-0ff40846f561\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.496222 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05df7d38-ed40-42b4-9761-0ff40846f561-serving-cert\") pod \"route-controller-manager-c99b4965f-q7mv9\" (UID: \"05df7d38-ed40-42b4-9761-0ff40846f561\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.496243 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spshl\" (UniqueName: \"kubernetes.io/projected/05df7d38-ed40-42b4-9761-0ff40846f561-kube-api-access-spshl\") pod \"route-controller-manager-c99b4965f-q7mv9\" (UID: \"05df7d38-ed40-42b4-9761-0ff40846f561\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.496277 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/05df7d38-ed40-42b4-9761-0ff40846f561-client-ca\") pod \"route-controller-manager-c99b4965f-q7mv9\" (UID: \"05df7d38-ed40-42b4-9761-0ff40846f561\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.597928 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05df7d38-ed40-42b4-9761-0ff40846f561-config\") pod \"route-controller-manager-c99b4965f-q7mv9\" (UID: \"05df7d38-ed40-42b4-9761-0ff40846f561\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.598270 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05df7d38-ed40-42b4-9761-0ff40846f561-serving-cert\") pod \"route-controller-manager-c99b4965f-q7mv9\" (UID: \"05df7d38-ed40-42b4-9761-0ff40846f561\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.598339 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spshl\" (UniqueName: \"kubernetes.io/projected/05df7d38-ed40-42b4-9761-0ff40846f561-kube-api-access-spshl\") pod \"route-controller-manager-c99b4965f-q7mv9\" (UID: \"05df7d38-ed40-42b4-9761-0ff40846f561\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.598443 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/05df7d38-ed40-42b4-9761-0ff40846f561-client-ca\") pod \"route-controller-manager-c99b4965f-q7mv9\" (UID: \"05df7d38-ed40-42b4-9761-0ff40846f561\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.599265 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/05df7d38-ed40-42b4-9761-0ff40846f561-client-ca\") pod \"route-controller-manager-c99b4965f-q7mv9\" (UID: \"05df7d38-ed40-42b4-9761-0ff40846f561\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.599500 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05df7d38-ed40-42b4-9761-0ff40846f561-config\") pod \"route-controller-manager-c99b4965f-q7mv9\" (UID: \"05df7d38-ed40-42b4-9761-0ff40846f561\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.603202 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05df7d38-ed40-42b4-9761-0ff40846f561-serving-cert\") pod \"route-controller-manager-c99b4965f-q7mv9\" (UID: \"05df7d38-ed40-42b4-9761-0ff40846f561\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.627043 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spshl\" (UniqueName: \"kubernetes.io/projected/05df7d38-ed40-42b4-9761-0ff40846f561-kube-api-access-spshl\") pod \"route-controller-manager-c99b4965f-q7mv9\" (UID: \"05df7d38-ed40-42b4-9761-0ff40846f561\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.685643 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.730751 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d1617e1-d79c-4197-ac74-edc70e60870d" path="/var/lib/kubelet/pods/7d1617e1-d79c-4197-ac74-edc70e60870d/volumes" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.933072 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9"] Feb 20 08:11:54 crc kubenswrapper[4948]: I0220 08:11:54.530692 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" event={"ID":"05df7d38-ed40-42b4-9761-0ff40846f561","Type":"ContainerStarted","Data":"4af19cbc8b5de82798fcfe4b39a57d6fcf401ac16b4e5756e3169eb3564d0f6f"} Feb 20 08:11:54 crc kubenswrapper[4948]: I0220 08:11:54.531114 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" event={"ID":"05df7d38-ed40-42b4-9761-0ff40846f561","Type":"ContainerStarted","Data":"423ce1385df7d56c79c46fdb192204a1248d92c9c0800a2af62972f7078e6f78"} Feb 20 08:11:54 crc kubenswrapper[4948]: I0220 08:11:54.531348 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" Feb 20 08:11:54 crc kubenswrapper[4948]: I0220 08:11:54.569170 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" podStartSLOduration=3.569140447 podStartE2EDuration="3.569140447s" podCreationTimestamp="2026-02-20 08:11:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:11:54.565155411 +0000 UTC m=+363.539650231" watchObservedRunningTime="2026-02-20 08:11:54.569140447 +0000 UTC m=+363.543635337" Feb 20 08:11:54 crc kubenswrapper[4948]: I0220 08:11:54.581616 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" Feb 20 08:12:07 crc kubenswrapper[4948]: I0220 08:12:07.215812 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:12:07 crc kubenswrapper[4948]: I0220 08:12:07.281185 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-7wm8c"] Feb 20 08:12:08 crc kubenswrapper[4948]: I0220 08:12:08.025194 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:12:08 crc kubenswrapper[4948]: I0220 08:12:08.025300 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.337303 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" podUID="57c455f7-aa9c-405c-bee3-89726b84f5db" containerName="registry" containerID="cri-o://20908318b48f1c9de151f5bb5f43a14fb60c5d750ea49471a9d574dabed15120" gracePeriod=30 Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.789932 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.790388 4948 generic.go:334] "Generic (PLEG): container finished" podID="57c455f7-aa9c-405c-bee3-89726b84f5db" containerID="20908318b48f1c9de151f5bb5f43a14fb60c5d750ea49471a9d574dabed15120" exitCode=0 Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.790437 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" event={"ID":"57c455f7-aa9c-405c-bee3-89726b84f5db","Type":"ContainerDied","Data":"20908318b48f1c9de151f5bb5f43a14fb60c5d750ea49471a9d574dabed15120"} Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.790494 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" event={"ID":"57c455f7-aa9c-405c-bee3-89726b84f5db","Type":"ContainerDied","Data":"f90694ba31dc9edaf32c68d5b113799c7c3cf39817bac7a06e98cc5e157c4d16"} Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.790517 4948 scope.go:117] "RemoveContainer" containerID="20908318b48f1c9de151f5bb5f43a14fb60c5d750ea49471a9d574dabed15120" Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.804599 4948 scope.go:117] "RemoveContainer" containerID="20908318b48f1c9de151f5bb5f43a14fb60c5d750ea49471a9d574dabed15120" Feb 20 08:12:32 crc kubenswrapper[4948]: E0220 08:12:32.805555 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20908318b48f1c9de151f5bb5f43a14fb60c5d750ea49471a9d574dabed15120\": container with ID starting with 20908318b48f1c9de151f5bb5f43a14fb60c5d750ea49471a9d574dabed15120 not found: ID does not exist" containerID="20908318b48f1c9de151f5bb5f43a14fb60c5d750ea49471a9d574dabed15120" Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.805625 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20908318b48f1c9de151f5bb5f43a14fb60c5d750ea49471a9d574dabed15120"} err="failed to get container status \"20908318b48f1c9de151f5bb5f43a14fb60c5d750ea49471a9d574dabed15120\": rpc error: code = NotFound desc = could not find container \"20908318b48f1c9de151f5bb5f43a14fb60c5d750ea49471a9d574dabed15120\": container with ID starting with 20908318b48f1c9de151f5bb5f43a14fb60c5d750ea49471a9d574dabed15120 not found: ID does not exist" Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.943353 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/57c455f7-aa9c-405c-bee3-89726b84f5db-installation-pull-secrets\") pod \"57c455f7-aa9c-405c-bee3-89726b84f5db\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.944140 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-bound-sa-token\") pod \"57c455f7-aa9c-405c-bee3-89726b84f5db\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.944211 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/57c455f7-aa9c-405c-bee3-89726b84f5db-trusted-ca\") pod \"57c455f7-aa9c-405c-bee3-89726b84f5db\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.944520 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"57c455f7-aa9c-405c-bee3-89726b84f5db\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.944562 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-registry-tls\") pod \"57c455f7-aa9c-405c-bee3-89726b84f5db\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.944608 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/57c455f7-aa9c-405c-bee3-89726b84f5db-registry-certificates\") pod \"57c455f7-aa9c-405c-bee3-89726b84f5db\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.945253 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2vg4v\" (UniqueName: \"kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-kube-api-access-2vg4v\") pod \"57c455f7-aa9c-405c-bee3-89726b84f5db\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.945357 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/57c455f7-aa9c-405c-bee3-89726b84f5db-ca-trust-extracted\") pod \"57c455f7-aa9c-405c-bee3-89726b84f5db\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.946246 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57c455f7-aa9c-405c-bee3-89726b84f5db-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "57c455f7-aa9c-405c-bee3-89726b84f5db" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.946450 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57c455f7-aa9c-405c-bee3-89726b84f5db-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "57c455f7-aa9c-405c-bee3-89726b84f5db" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.956631 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-kube-api-access-2vg4v" (OuterVolumeSpecName: "kube-api-access-2vg4v") pod "57c455f7-aa9c-405c-bee3-89726b84f5db" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db"). InnerVolumeSpecName "kube-api-access-2vg4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.957359 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57c455f7-aa9c-405c-bee3-89726b84f5db-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "57c455f7-aa9c-405c-bee3-89726b84f5db" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.957888 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "57c455f7-aa9c-405c-bee3-89726b84f5db" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.957991 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "57c455f7-aa9c-405c-bee3-89726b84f5db" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.958335 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "57c455f7-aa9c-405c-bee3-89726b84f5db" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.983810 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57c455f7-aa9c-405c-bee3-89726b84f5db-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "57c455f7-aa9c-405c-bee3-89726b84f5db" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:12:33 crc kubenswrapper[4948]: I0220 08:12:33.047128 4948 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/57c455f7-aa9c-405c-bee3-89726b84f5db-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:12:33 crc kubenswrapper[4948]: I0220 08:12:33.047195 4948 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 20 08:12:33 crc kubenswrapper[4948]: I0220 08:12:33.047224 4948 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/57c455f7-aa9c-405c-bee3-89726b84f5db-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 20 08:12:33 crc kubenswrapper[4948]: I0220 08:12:33.047251 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2vg4v\" (UniqueName: \"kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-kube-api-access-2vg4v\") on node \"crc\" DevicePath \"\"" Feb 20 08:12:33 crc kubenswrapper[4948]: I0220 08:12:33.047274 4948 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/57c455f7-aa9c-405c-bee3-89726b84f5db-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 20 08:12:33 crc kubenswrapper[4948]: I0220 08:12:33.047298 4948 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/57c455f7-aa9c-405c-bee3-89726b84f5db-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 20 08:12:33 crc kubenswrapper[4948]: I0220 08:12:33.047321 4948 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 20 08:12:33 crc kubenswrapper[4948]: I0220 08:12:33.797725 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:12:33 crc kubenswrapper[4948]: I0220 08:12:33.822381 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-7wm8c"] Feb 20 08:12:33 crc kubenswrapper[4948]: I0220 08:12:33.826664 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-7wm8c"] Feb 20 08:12:35 crc kubenswrapper[4948]: I0220 08:12:35.735478 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57c455f7-aa9c-405c-bee3-89726b84f5db" path="/var/lib/kubelet/pods/57c455f7-aa9c-405c-bee3-89726b84f5db/volumes" Feb 20 08:12:37 crc kubenswrapper[4948]: I0220 08:12:37.597307 4948 patch_prober.go:28] interesting pod/image-registry-697d97f7c8-7wm8c container/registry namespace/openshift-image-registry: Readiness probe status=failure output="Get \"https://10.217.0.20:5000/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 20 08:12:37 crc kubenswrapper[4948]: I0220 08:12:37.598229 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" podUID="57c455f7-aa9c-405c-bee3-89726b84f5db" containerName="registry" probeResult="failure" output="Get \"https://10.217.0.20:5000/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 20 08:12:38 crc kubenswrapper[4948]: I0220 08:12:38.024531 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:12:38 crc kubenswrapper[4948]: I0220 08:12:38.024656 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:12:38 crc kubenswrapper[4948]: I0220 08:12:38.024717 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:12:38 crc kubenswrapper[4948]: I0220 08:12:38.025561 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6791802a782456e3c45df951fa55097ce0ea3affd6a9878fc2981511046d1b09"} pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 08:12:38 crc kubenswrapper[4948]: I0220 08:12:38.025705 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" containerID="cri-o://6791802a782456e3c45df951fa55097ce0ea3affd6a9878fc2981511046d1b09" gracePeriod=600 Feb 20 08:12:38 crc kubenswrapper[4948]: I0220 08:12:38.834144 4948 generic.go:334] "Generic (PLEG): container finished" podID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerID="6791802a782456e3c45df951fa55097ce0ea3affd6a9878fc2981511046d1b09" exitCode=0 Feb 20 08:12:38 crc kubenswrapper[4948]: I0220 08:12:38.834200 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerDied","Data":"6791802a782456e3c45df951fa55097ce0ea3affd6a9878fc2981511046d1b09"} Feb 20 08:12:38 crc kubenswrapper[4948]: I0220 08:12:38.834619 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"a2c247070dfb4d0488751bb46e5543e834b3853c983d98f00e419c6aabfdeb7c"} Feb 20 08:12:38 crc kubenswrapper[4948]: I0220 08:12:38.834642 4948 scope.go:117] "RemoveContainer" containerID="874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59" Feb 20 08:14:38 crc kubenswrapper[4948]: I0220 08:14:38.024735 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:14:38 crc kubenswrapper[4948]: I0220 08:14:38.025344 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.203298 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9"] Feb 20 08:15:00 crc kubenswrapper[4948]: E0220 08:15:00.204510 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57c455f7-aa9c-405c-bee3-89726b84f5db" containerName="registry" Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.204540 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="57c455f7-aa9c-405c-bee3-89726b84f5db" containerName="registry" Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.204779 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="57c455f7-aa9c-405c-bee3-89726b84f5db" containerName="registry" Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.205623 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9" Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.211357 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.212172 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.218276 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9"] Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.384581 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9db735d8-ae70-471b-9d86-3be100901c72-secret-volume\") pod \"collect-profiles-29526255-62px9\" (UID: \"9db735d8-ae70-471b-9d86-3be100901c72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9" Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.384754 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqj9z\" (UniqueName: \"kubernetes.io/projected/9db735d8-ae70-471b-9d86-3be100901c72-kube-api-access-fqj9z\") pod \"collect-profiles-29526255-62px9\" (UID: \"9db735d8-ae70-471b-9d86-3be100901c72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9" Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.384865 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9db735d8-ae70-471b-9d86-3be100901c72-config-volume\") pod \"collect-profiles-29526255-62px9\" (UID: \"9db735d8-ae70-471b-9d86-3be100901c72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9" Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.485661 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9db735d8-ae70-471b-9d86-3be100901c72-secret-volume\") pod \"collect-profiles-29526255-62px9\" (UID: \"9db735d8-ae70-471b-9d86-3be100901c72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9" Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.485738 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqj9z\" (UniqueName: \"kubernetes.io/projected/9db735d8-ae70-471b-9d86-3be100901c72-kube-api-access-fqj9z\") pod \"collect-profiles-29526255-62px9\" (UID: \"9db735d8-ae70-471b-9d86-3be100901c72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9" Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.485774 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9db735d8-ae70-471b-9d86-3be100901c72-config-volume\") pod \"collect-profiles-29526255-62px9\" (UID: \"9db735d8-ae70-471b-9d86-3be100901c72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9" Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.486740 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9db735d8-ae70-471b-9d86-3be100901c72-config-volume\") pod \"collect-profiles-29526255-62px9\" (UID: \"9db735d8-ae70-471b-9d86-3be100901c72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9" Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.495368 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9db735d8-ae70-471b-9d86-3be100901c72-secret-volume\") pod \"collect-profiles-29526255-62px9\" (UID: \"9db735d8-ae70-471b-9d86-3be100901c72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9" Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.508321 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqj9z\" (UniqueName: \"kubernetes.io/projected/9db735d8-ae70-471b-9d86-3be100901c72-kube-api-access-fqj9z\") pod \"collect-profiles-29526255-62px9\" (UID: \"9db735d8-ae70-471b-9d86-3be100901c72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9" Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.538318 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9" Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.741669 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9"] Feb 20 08:15:00 crc kubenswrapper[4948]: W0220 08:15:00.752081 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9db735d8_ae70_471b_9d86_3be100901c72.slice/crio-0801fcdfdc4f8abf5fd4ba611ffdd02a9bfe2e69c820029184f902432874be1d WatchSource:0}: Error finding container 0801fcdfdc4f8abf5fd4ba611ffdd02a9bfe2e69c820029184f902432874be1d: Status 404 returned error can't find the container with id 0801fcdfdc4f8abf5fd4ba611ffdd02a9bfe2e69c820029184f902432874be1d Feb 20 08:15:01 crc kubenswrapper[4948]: I0220 08:15:01.332478 4948 generic.go:334] "Generic (PLEG): container finished" podID="9db735d8-ae70-471b-9d86-3be100901c72" containerID="984bcec5b88d4682af52feec68d576f83622efc6d9578f02504c68786eff5cfd" exitCode=0 Feb 20 08:15:01 crc kubenswrapper[4948]: I0220 08:15:01.332733 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9" event={"ID":"9db735d8-ae70-471b-9d86-3be100901c72","Type":"ContainerDied","Data":"984bcec5b88d4682af52feec68d576f83622efc6d9578f02504c68786eff5cfd"} Feb 20 08:15:01 crc kubenswrapper[4948]: I0220 08:15:01.333015 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9" event={"ID":"9db735d8-ae70-471b-9d86-3be100901c72","Type":"ContainerStarted","Data":"0801fcdfdc4f8abf5fd4ba611ffdd02a9bfe2e69c820029184f902432874be1d"} Feb 20 08:15:02 crc kubenswrapper[4948]: I0220 08:15:02.606279 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9" Feb 20 08:15:02 crc kubenswrapper[4948]: I0220 08:15:02.715962 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9db735d8-ae70-471b-9d86-3be100901c72-secret-volume\") pod \"9db735d8-ae70-471b-9d86-3be100901c72\" (UID: \"9db735d8-ae70-471b-9d86-3be100901c72\") " Feb 20 08:15:02 crc kubenswrapper[4948]: I0220 08:15:02.716097 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9db735d8-ae70-471b-9d86-3be100901c72-config-volume\") pod \"9db735d8-ae70-471b-9d86-3be100901c72\" (UID: \"9db735d8-ae70-471b-9d86-3be100901c72\") " Feb 20 08:15:02 crc kubenswrapper[4948]: I0220 08:15:02.716173 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqj9z\" (UniqueName: \"kubernetes.io/projected/9db735d8-ae70-471b-9d86-3be100901c72-kube-api-access-fqj9z\") pod \"9db735d8-ae70-471b-9d86-3be100901c72\" (UID: \"9db735d8-ae70-471b-9d86-3be100901c72\") " Feb 20 08:15:02 crc kubenswrapper[4948]: I0220 08:15:02.717182 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9db735d8-ae70-471b-9d86-3be100901c72-config-volume" (OuterVolumeSpecName: "config-volume") pod "9db735d8-ae70-471b-9d86-3be100901c72" (UID: "9db735d8-ae70-471b-9d86-3be100901c72"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:15:02 crc kubenswrapper[4948]: I0220 08:15:02.722738 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9db735d8-ae70-471b-9d86-3be100901c72-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9db735d8-ae70-471b-9d86-3be100901c72" (UID: "9db735d8-ae70-471b-9d86-3be100901c72"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:15:02 crc kubenswrapper[4948]: I0220 08:15:02.723631 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9db735d8-ae70-471b-9d86-3be100901c72-kube-api-access-fqj9z" (OuterVolumeSpecName: "kube-api-access-fqj9z") pod "9db735d8-ae70-471b-9d86-3be100901c72" (UID: "9db735d8-ae70-471b-9d86-3be100901c72"). InnerVolumeSpecName "kube-api-access-fqj9z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:15:02 crc kubenswrapper[4948]: I0220 08:15:02.818261 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqj9z\" (UniqueName: \"kubernetes.io/projected/9db735d8-ae70-471b-9d86-3be100901c72-kube-api-access-fqj9z\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:02 crc kubenswrapper[4948]: I0220 08:15:02.818362 4948 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9db735d8-ae70-471b-9d86-3be100901c72-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:02 crc kubenswrapper[4948]: I0220 08:15:02.818391 4948 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9db735d8-ae70-471b-9d86-3be100901c72-config-volume\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:03 crc kubenswrapper[4948]: I0220 08:15:03.349623 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9" event={"ID":"9db735d8-ae70-471b-9d86-3be100901c72","Type":"ContainerDied","Data":"0801fcdfdc4f8abf5fd4ba611ffdd02a9bfe2e69c820029184f902432874be1d"} Feb 20 08:15:03 crc kubenswrapper[4948]: I0220 08:15:03.350443 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0801fcdfdc4f8abf5fd4ba611ffdd02a9bfe2e69c820029184f902432874be1d" Feb 20 08:15:03 crc kubenswrapper[4948]: I0220 08:15:03.350402 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9" Feb 20 08:15:08 crc kubenswrapper[4948]: I0220 08:15:08.024683 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:15:08 crc kubenswrapper[4948]: I0220 08:15:08.024746 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.381988 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-9llfv"] Feb 20 08:15:24 crc kubenswrapper[4948]: E0220 08:15:24.384754 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9db735d8-ae70-471b-9d86-3be100901c72" containerName="collect-profiles" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.384771 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="9db735d8-ae70-471b-9d86-3be100901c72" containerName="collect-profiles" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.386343 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="9db735d8-ae70-471b-9d86-3be100901c72" containerName="collect-profiles" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.387179 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-9llfv" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.390601 4948 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-k2zw5" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.392185 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.392230 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.392470 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-smtqw"] Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.393525 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-smtqw" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.394927 4948 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-v2wrp" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.401125 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-9llfv"] Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.424237 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-smtqw"] Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.436856 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-f4bjc"] Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.437693 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-f4bjc" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.441523 4948 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-jqmnm" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.450434 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-f4bjc"] Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.552832 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftvfl\" (UniqueName: \"kubernetes.io/projected/bb56399a-a840-49af-972f-4f32c91efed7-kube-api-access-ftvfl\") pod \"cert-manager-cainjector-cf98fcc89-9llfv\" (UID: \"bb56399a-a840-49af-972f-4f32c91efed7\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-9llfv" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.552973 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8h75\" (UniqueName: \"kubernetes.io/projected/305a17f4-aef3-4036-8fce-3756ff5bbd2f-kube-api-access-l8h75\") pod \"cert-manager-webhook-687f57d79b-f4bjc\" (UID: \"305a17f4-aef3-4036-8fce-3756ff5bbd2f\") " pod="cert-manager/cert-manager-webhook-687f57d79b-f4bjc" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.553064 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jrjz\" (UniqueName: \"kubernetes.io/projected/1437d818-abf4-4602-8a28-e88a76e482a9-kube-api-access-9jrjz\") pod \"cert-manager-858654f9db-smtqw\" (UID: \"1437d818-abf4-4602-8a28-e88a76e482a9\") " pod="cert-manager/cert-manager-858654f9db-smtqw" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.654136 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8h75\" (UniqueName: \"kubernetes.io/projected/305a17f4-aef3-4036-8fce-3756ff5bbd2f-kube-api-access-l8h75\") pod \"cert-manager-webhook-687f57d79b-f4bjc\" (UID: \"305a17f4-aef3-4036-8fce-3756ff5bbd2f\") " pod="cert-manager/cert-manager-webhook-687f57d79b-f4bjc" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.654238 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jrjz\" (UniqueName: \"kubernetes.io/projected/1437d818-abf4-4602-8a28-e88a76e482a9-kube-api-access-9jrjz\") pod \"cert-manager-858654f9db-smtqw\" (UID: \"1437d818-abf4-4602-8a28-e88a76e482a9\") " pod="cert-manager/cert-manager-858654f9db-smtqw" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.654293 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftvfl\" (UniqueName: \"kubernetes.io/projected/bb56399a-a840-49af-972f-4f32c91efed7-kube-api-access-ftvfl\") pod \"cert-manager-cainjector-cf98fcc89-9llfv\" (UID: \"bb56399a-a840-49af-972f-4f32c91efed7\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-9llfv" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.677920 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8h75\" (UniqueName: \"kubernetes.io/projected/305a17f4-aef3-4036-8fce-3756ff5bbd2f-kube-api-access-l8h75\") pod \"cert-manager-webhook-687f57d79b-f4bjc\" (UID: \"305a17f4-aef3-4036-8fce-3756ff5bbd2f\") " pod="cert-manager/cert-manager-webhook-687f57d79b-f4bjc" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.678590 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftvfl\" (UniqueName: \"kubernetes.io/projected/bb56399a-a840-49af-972f-4f32c91efed7-kube-api-access-ftvfl\") pod \"cert-manager-cainjector-cf98fcc89-9llfv\" (UID: \"bb56399a-a840-49af-972f-4f32c91efed7\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-9llfv" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.681000 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jrjz\" (UniqueName: \"kubernetes.io/projected/1437d818-abf4-4602-8a28-e88a76e482a9-kube-api-access-9jrjz\") pod \"cert-manager-858654f9db-smtqw\" (UID: \"1437d818-abf4-4602-8a28-e88a76e482a9\") " pod="cert-manager/cert-manager-858654f9db-smtqw" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.704900 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-9llfv" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.712165 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-smtqw" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.751043 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-f4bjc" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.983671 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-f4bjc"] Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.999325 4948 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 08:15:25 crc kubenswrapper[4948]: I0220 08:15:25.132635 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-smtqw"] Feb 20 08:15:25 crc kubenswrapper[4948]: I0220 08:15:25.149846 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-9llfv"] Feb 20 08:15:25 crc kubenswrapper[4948]: W0220 08:15:25.156206 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb56399a_a840_49af_972f_4f32c91efed7.slice/crio-d0eaef1010af91c5fcc7aaadb45551119645a38b302145a9450750c0da2c7c34 WatchSource:0}: Error finding container d0eaef1010af91c5fcc7aaadb45551119645a38b302145a9450750c0da2c7c34: Status 404 returned error can't find the container with id d0eaef1010af91c5fcc7aaadb45551119645a38b302145a9450750c0da2c7c34 Feb 20 08:15:25 crc kubenswrapper[4948]: I0220 08:15:25.501963 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-f4bjc" event={"ID":"305a17f4-aef3-4036-8fce-3756ff5bbd2f","Type":"ContainerStarted","Data":"32bc60676f3e532776c6c21ba8be09d090b22873a836263a5e246756ccba5bb9"} Feb 20 08:15:25 crc kubenswrapper[4948]: I0220 08:15:25.503436 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-smtqw" event={"ID":"1437d818-abf4-4602-8a28-e88a76e482a9","Type":"ContainerStarted","Data":"fbaf331b99a09cd4c0955dfc5844f6e128026c3e9782f085962e44bd1d0ab81b"} Feb 20 08:15:25 crc kubenswrapper[4948]: I0220 08:15:25.505221 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-9llfv" event={"ID":"bb56399a-a840-49af-972f-4f32c91efed7","Type":"ContainerStarted","Data":"d0eaef1010af91c5fcc7aaadb45551119645a38b302145a9450750c0da2c7c34"} Feb 20 08:15:30 crc kubenswrapper[4948]: I0220 08:15:30.542185 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-smtqw" event={"ID":"1437d818-abf4-4602-8a28-e88a76e482a9","Type":"ContainerStarted","Data":"f3c8fec8b780bb055690941032243c45f7653a3e6b93dffb13ac54496d436e7e"} Feb 20 08:15:30 crc kubenswrapper[4948]: I0220 08:15:30.544133 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-9llfv" event={"ID":"bb56399a-a840-49af-972f-4f32c91efed7","Type":"ContainerStarted","Data":"552830168d2299c90df97b8c06eb074c90da6db5087c1bcf8fd36c3079dbacc4"} Feb 20 08:15:30 crc kubenswrapper[4948]: I0220 08:15:30.545984 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-f4bjc" event={"ID":"305a17f4-aef3-4036-8fce-3756ff5bbd2f","Type":"ContainerStarted","Data":"a8d19f39147e5e961003b5aaac53e6b1e65d2c280885386da9cc14e15b28ad8a"} Feb 20 08:15:30 crc kubenswrapper[4948]: I0220 08:15:30.546227 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-f4bjc" Feb 20 08:15:30 crc kubenswrapper[4948]: I0220 08:15:30.570848 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-smtqw" podStartSLOduration=2.39105744 podStartE2EDuration="6.570818521s" podCreationTimestamp="2026-02-20 08:15:24 +0000 UTC" firstStartedPulling="2026-02-20 08:15:25.141489303 +0000 UTC m=+574.115984123" lastFinishedPulling="2026-02-20 08:15:29.321250374 +0000 UTC m=+578.295745204" observedRunningTime="2026-02-20 08:15:30.564042273 +0000 UTC m=+579.538537113" watchObservedRunningTime="2026-02-20 08:15:30.570818521 +0000 UTC m=+579.545313361" Feb 20 08:15:30 crc kubenswrapper[4948]: I0220 08:15:30.581205 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-9llfv" podStartSLOduration=2.4194332960000002 podStartE2EDuration="6.581177289s" podCreationTimestamp="2026-02-20 08:15:24 +0000 UTC" firstStartedPulling="2026-02-20 08:15:25.159793408 +0000 UTC m=+574.134288228" lastFinishedPulling="2026-02-20 08:15:29.321537361 +0000 UTC m=+578.296032221" observedRunningTime="2026-02-20 08:15:30.58044695 +0000 UTC m=+579.554941780" watchObservedRunningTime="2026-02-20 08:15:30.581177289 +0000 UTC m=+579.555672129" Feb 20 08:15:30 crc kubenswrapper[4948]: I0220 08:15:30.618850 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-f4bjc" podStartSLOduration=2.211713786 podStartE2EDuration="6.618823624s" podCreationTimestamp="2026-02-20 08:15:24 +0000 UTC" firstStartedPulling="2026-02-20 08:15:24.999104577 +0000 UTC m=+573.973599397" lastFinishedPulling="2026-02-20 08:15:29.406214405 +0000 UTC m=+578.380709235" observedRunningTime="2026-02-20 08:15:30.61584189 +0000 UTC m=+579.590336750" watchObservedRunningTime="2026-02-20 08:15:30.618823624 +0000 UTC m=+579.593318454" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.622433 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7s8gg"] Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.623772 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovn-controller" containerID="cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d" gracePeriod=30 Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.624004 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="northd" containerID="cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2" gracePeriod=30 Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.624171 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37" gracePeriod=30 Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.624319 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="kube-rbac-proxy-node" containerID="cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c" gracePeriod=30 Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.624357 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovn-acl-logging" containerID="cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7" gracePeriod=30 Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.624510 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="sbdb" containerID="cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022" gracePeriod=30 Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.624928 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="nbdb" containerID="cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a" gracePeriod=30 Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.665753 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovnkube-controller" containerID="cri-o://682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a" gracePeriod=30 Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.754358 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-f4bjc" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.907391 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovnkube-controller/3.log" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.910426 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovn-acl-logging/0.log" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.911332 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovn-controller/0.log" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.912068 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.960418 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-cvhbs"] Feb 20 08:15:34 crc kubenswrapper[4948]: E0220 08:15:34.960643 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="kube-rbac-proxy-node" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.960655 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="kube-rbac-proxy-node" Feb 20 08:15:34 crc kubenswrapper[4948]: E0220 08:15:34.960669 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovn-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.960676 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovn-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: E0220 08:15:34.960684 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovn-acl-logging" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.960690 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovn-acl-logging" Feb 20 08:15:34 crc kubenswrapper[4948]: E0220 08:15:34.960698 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovnkube-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.960705 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovnkube-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: E0220 08:15:34.960712 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="kubecfg-setup" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.960719 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="kubecfg-setup" Feb 20 08:15:34 crc kubenswrapper[4948]: E0220 08:15:34.960728 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="nbdb" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.960734 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="nbdb" Feb 20 08:15:34 crc kubenswrapper[4948]: E0220 08:15:34.960745 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovnkube-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.960753 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovnkube-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: E0220 08:15:34.960764 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovnkube-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.960771 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovnkube-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: E0220 08:15:34.960781 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="sbdb" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.960788 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="sbdb" Feb 20 08:15:34 crc kubenswrapper[4948]: E0220 08:15:34.960797 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovnkube-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.960805 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovnkube-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: E0220 08:15:34.960814 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="kube-rbac-proxy-ovn-metrics" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.960821 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="kube-rbac-proxy-ovn-metrics" Feb 20 08:15:34 crc kubenswrapper[4948]: E0220 08:15:34.960833 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="northd" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.960839 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="northd" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.960945 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovnkube-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.960959 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="sbdb" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.960993 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovnkube-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.961004 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="kube-rbac-proxy-ovn-metrics" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.961013 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovn-acl-logging" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.961022 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovnkube-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.961030 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovnkube-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.961040 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovnkube-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.961050 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovn-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.961061 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="northd" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.961071 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="kube-rbac-proxy-node" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.961080 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="nbdb" Feb 20 08:15:34 crc kubenswrapper[4948]: E0220 08:15:34.961368 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovnkube-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.961377 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovnkube-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.963213 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.043874 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovnkube-config\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.044283 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-openvswitch\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.044328 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-env-overrides\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.044368 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-etc-openvswitch\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.044397 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-run-netns\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.044416 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.044430 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-systemd-units\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.044479 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.044491 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovnkube-script-lib\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.044511 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-ovn\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.044531 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.044567 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.044601 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.044704 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-slash\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.044726 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-slash" (OuterVolumeSpecName: "host-slash") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.044798 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.044921 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-node-log\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.044962 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovn-node-metrics-cert\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045006 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045024 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-kubelet\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045036 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045060 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-systemd\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045050 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-node-log" (OuterVolumeSpecName: "node-log") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045099 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-cni-netd\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045106 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045136 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-cni-bin\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045162 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-var-lib-cni-networks-ovn-kubernetes\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045188 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtrs5\" (UniqueName: \"kubernetes.io/projected/d6a28bd7-2b83-43f8-b803-bfe41516e071-kube-api-access-gtrs5\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045211 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-var-lib-openvswitch\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045224 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045229 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-run-ovn-kubernetes\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045250 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045254 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-log-socket\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045283 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-log-socket" (OuterVolumeSpecName: "log-socket") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045315 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045329 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045365 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-run-netns\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045393 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-cni-bin\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045411 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-cni-netd\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045424 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0bfc877e-59ed-4803-b07a-b9787b1ab73d-ovnkube-script-lib\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045443 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-kubelet\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045465 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-run-openvswitch\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045494 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-systemd-units\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045512 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0bfc877e-59ed-4803-b07a-b9787b1ab73d-ovnkube-config\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045526 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0bfc877e-59ed-4803-b07a-b9787b1ab73d-ovn-node-metrics-cert\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045551 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsmb9\" (UniqueName: \"kubernetes.io/projected/0bfc877e-59ed-4803-b07a-b9787b1ab73d-kube-api-access-lsmb9\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045581 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-node-log\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045611 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-run-systemd\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045628 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-run-ovn\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045643 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-run-ovn-kubernetes\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045669 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-etc-openvswitch\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045685 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0bfc877e-59ed-4803-b07a-b9787b1ab73d-env-overrides\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045700 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-slash\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045715 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-var-lib-openvswitch\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045745 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-log-socket\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045783 4948 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045793 4948 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045802 4948 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045810 4948 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-run-netns\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045818 4948 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-systemd-units\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045826 4948 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045834 4948 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045842 4948 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-slash\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045852 4948 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-node-log\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045841 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045860 4948 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-kubelet\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045928 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.046066 4948 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-cni-netd\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.046100 4948 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-cni-bin\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.046123 4948 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.046142 4948 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-log-socket\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.046159 4948 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.053316 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6a28bd7-2b83-43f8-b803-bfe41516e071-kube-api-access-gtrs5" (OuterVolumeSpecName: "kube-api-access-gtrs5") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "kube-api-access-gtrs5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.053322 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.071952 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.146875 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-etc-openvswitch\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.146940 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0bfc877e-59ed-4803-b07a-b9787b1ab73d-env-overrides\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.146993 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-slash\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147024 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-var-lib-openvswitch\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147070 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-log-socket\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147065 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-etc-openvswitch\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147109 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147173 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147280 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-log-socket\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147232 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-var-lib-openvswitch\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147233 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-run-netns\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147285 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-run-netns\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147443 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-cni-bin\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147181 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-slash\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147500 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-cni-netd\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147550 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-cni-netd\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147560 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-cni-bin\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147572 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0bfc877e-59ed-4803-b07a-b9787b1ab73d-ovnkube-script-lib\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147660 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-kubelet\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147720 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-run-openvswitch\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147743 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-kubelet\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147789 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-systemd-units\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147837 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0bfc877e-59ed-4803-b07a-b9787b1ab73d-ovnkube-config\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147880 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0bfc877e-59ed-4803-b07a-b9787b1ab73d-ovn-node-metrics-cert\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147889 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-systemd-units\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147938 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsmb9\" (UniqueName: \"kubernetes.io/projected/0bfc877e-59ed-4803-b07a-b9787b1ab73d-kube-api-access-lsmb9\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147832 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-run-openvswitch\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.148047 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-node-log\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.148146 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-node-log\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.148318 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-run-systemd\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.148373 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-run-ovn\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.148413 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-run-ovn-kubernetes\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.148461 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-run-systemd\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.148514 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-run-ovn\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.148595 4948 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.148601 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-run-ovn-kubernetes\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.148610 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0bfc877e-59ed-4803-b07a-b9787b1ab73d-env-overrides\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.148618 4948 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-systemd\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.148708 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtrs5\" (UniqueName: \"kubernetes.io/projected/d6a28bd7-2b83-43f8-b803-bfe41516e071-kube-api-access-gtrs5\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.148728 4948 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.148747 4948 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.148774 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0bfc877e-59ed-4803-b07a-b9787b1ab73d-ovnkube-script-lib\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.149672 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0bfc877e-59ed-4803-b07a-b9787b1ab73d-ovnkube-config\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.153614 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0bfc877e-59ed-4803-b07a-b9787b1ab73d-ovn-node-metrics-cert\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.176899 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsmb9\" (UniqueName: \"kubernetes.io/projected/0bfc877e-59ed-4803-b07a-b9787b1ab73d-kube-api-access-lsmb9\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.276309 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: W0220 08:15:35.313065 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0bfc877e_59ed_4803_b07a_b9787b1ab73d.slice/crio-b2bcab2b0e6efcd12fe28db5042d25c8f825151e8a2bf70419baa020ac3e634d WatchSource:0}: Error finding container b2bcab2b0e6efcd12fe28db5042d25c8f825151e8a2bf70419baa020ac3e634d: Status 404 returned error can't find the container with id b2bcab2b0e6efcd12fe28db5042d25c8f825151e8a2bf70419baa020ac3e634d Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.586106 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovnkube-controller/3.log" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.590408 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovn-acl-logging/0.log" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.591284 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovn-controller/0.log" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.592019 4948 generic.go:334] "Generic (PLEG): container finished" podID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerID="682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a" exitCode=0 Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.592168 4948 generic.go:334] "Generic (PLEG): container finished" podID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerID="c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022" exitCode=0 Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.592304 4948 generic.go:334] "Generic (PLEG): container finished" podID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerID="8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a" exitCode=0 Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.592417 4948 generic.go:334] "Generic (PLEG): container finished" podID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerID="5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2" exitCode=0 Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.592538 4948 generic.go:334] "Generic (PLEG): container finished" podID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerID="6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37" exitCode=0 Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.592653 4948 generic.go:334] "Generic (PLEG): container finished" podID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerID="719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c" exitCode=0 Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.592754 4948 generic.go:334] "Generic (PLEG): container finished" podID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerID="3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7" exitCode=143 Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.592490 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.592895 4948 generic.go:334] "Generic (PLEG): container finished" podID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerID="9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d" exitCode=143 Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.592333 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerDied","Data":"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.593494 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerDied","Data":"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.593666 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerDied","Data":"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.593786 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerDied","Data":"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.593920 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerDied","Data":"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.594074 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerDied","Data":"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.594217 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.594336 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.594442 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.594539 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.594642 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.594746 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.594857 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.594966 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595101 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595224 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerDied","Data":"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595343 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595460 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595479 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595489 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595502 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595513 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595524 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595534 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595545 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595555 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595572 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerDied","Data":"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595617 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595631 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595642 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595653 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595663 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595674 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595684 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595695 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595705 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595715 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595729 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerDied","Data":"0269d0e4618f9bfc08bbd4b396cc89a6bd78be17afe5c9425cf43f02da885bc5"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595745 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595758 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595771 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595781 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595792 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595803 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595813 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595823 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595834 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595845 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.593839 4948 scope.go:117] "RemoveContainer" containerID="682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.603693 4948 generic.go:334] "Generic (PLEG): container finished" podID="0bfc877e-59ed-4803-b07a-b9787b1ab73d" containerID="83c11a6ecf91713d5e31e2539b1a573dbcda58bd23320b28bcb8d4ab32147fb4" exitCode=0 Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.603795 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" event={"ID":"0bfc877e-59ed-4803-b07a-b9787b1ab73d","Type":"ContainerDied","Data":"83c11a6ecf91713d5e31e2539b1a573dbcda58bd23320b28bcb8d4ab32147fb4"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.603841 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" event={"ID":"0bfc877e-59ed-4803-b07a-b9787b1ab73d","Type":"ContainerStarted","Data":"b2bcab2b0e6efcd12fe28db5042d25c8f825151e8a2bf70419baa020ac3e634d"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.613690 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x7s56_b96124e4-0a74-4578-9142-fd728eb9f99e/kube-multus/2.log" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.614454 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x7s56_b96124e4-0a74-4578-9142-fd728eb9f99e/kube-multus/1.log" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.614535 4948 generic.go:334] "Generic (PLEG): container finished" podID="b96124e4-0a74-4578-9142-fd728eb9f99e" containerID="4ecf8c7c9ed34410ff4619b18254e2eb90170911300e6cd6a7284e84e8937246" exitCode=2 Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.614581 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x7s56" event={"ID":"b96124e4-0a74-4578-9142-fd728eb9f99e","Type":"ContainerDied","Data":"4ecf8c7c9ed34410ff4619b18254e2eb90170911300e6cd6a7284e84e8937246"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.614650 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f4093ec9d71181c58ad80a8c4ee3692148590df14aadaec96c0f54a843d1f8c"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.615432 4948 scope.go:117] "RemoveContainer" containerID="4ecf8c7c9ed34410ff4619b18254e2eb90170911300e6cd6a7284e84e8937246" Feb 20 08:15:35 crc kubenswrapper[4948]: E0220 08:15:35.615781 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-x7s56_openshift-multus(b96124e4-0a74-4578-9142-fd728eb9f99e)\"" pod="openshift-multus/multus-x7s56" podUID="b96124e4-0a74-4578-9142-fd728eb9f99e" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.631042 4948 scope.go:117] "RemoveContainer" containerID="7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.691850 4948 scope.go:117] "RemoveContainer" containerID="c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.699640 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7s8gg"] Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.706844 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7s8gg"] Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.719393 4948 scope.go:117] "RemoveContainer" containerID="8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.732012 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" path="/var/lib/kubelet/pods/d6a28bd7-2b83-43f8-b803-bfe41516e071/volumes" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.743312 4948 scope.go:117] "RemoveContainer" containerID="5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.763432 4948 scope.go:117] "RemoveContainer" containerID="6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.799846 4948 scope.go:117] "RemoveContainer" containerID="719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.814893 4948 scope.go:117] "RemoveContainer" containerID="3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.830619 4948 scope.go:117] "RemoveContainer" containerID="9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.859531 4948 scope.go:117] "RemoveContainer" containerID="540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.893047 4948 scope.go:117] "RemoveContainer" containerID="682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a" Feb 20 08:15:35 crc kubenswrapper[4948]: E0220 08:15:35.894633 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a\": container with ID starting with 682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a not found: ID does not exist" containerID="682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.894689 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a"} err="failed to get container status \"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a\": rpc error: code = NotFound desc = could not find container \"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a\": container with ID starting with 682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.894727 4948 scope.go:117] "RemoveContainer" containerID="7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b" Feb 20 08:15:35 crc kubenswrapper[4948]: E0220 08:15:35.895456 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b\": container with ID starting with 7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b not found: ID does not exist" containerID="7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.895508 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b"} err="failed to get container status \"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b\": rpc error: code = NotFound desc = could not find container \"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b\": container with ID starting with 7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.895542 4948 scope.go:117] "RemoveContainer" containerID="c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022" Feb 20 08:15:35 crc kubenswrapper[4948]: E0220 08:15:35.895923 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\": container with ID starting with c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022 not found: ID does not exist" containerID="c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.895960 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022"} err="failed to get container status \"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\": rpc error: code = NotFound desc = could not find container \"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\": container with ID starting with c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022 not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.896008 4948 scope.go:117] "RemoveContainer" containerID="8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a" Feb 20 08:15:35 crc kubenswrapper[4948]: E0220 08:15:35.896863 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\": container with ID starting with 8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a not found: ID does not exist" containerID="8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.896948 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a"} err="failed to get container status \"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\": rpc error: code = NotFound desc = could not find container \"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\": container with ID starting with 8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.897099 4948 scope.go:117] "RemoveContainer" containerID="5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2" Feb 20 08:15:35 crc kubenswrapper[4948]: E0220 08:15:35.898565 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\": container with ID starting with 5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2 not found: ID does not exist" containerID="5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.898714 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2"} err="failed to get container status \"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\": rpc error: code = NotFound desc = could not find container \"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\": container with ID starting with 5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2 not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.898755 4948 scope.go:117] "RemoveContainer" containerID="6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37" Feb 20 08:15:35 crc kubenswrapper[4948]: E0220 08:15:35.899405 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\": container with ID starting with 6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37 not found: ID does not exist" containerID="6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.899444 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37"} err="failed to get container status \"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\": rpc error: code = NotFound desc = could not find container \"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\": container with ID starting with 6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37 not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.899472 4948 scope.go:117] "RemoveContainer" containerID="719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c" Feb 20 08:15:35 crc kubenswrapper[4948]: E0220 08:15:35.899772 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\": container with ID starting with 719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c not found: ID does not exist" containerID="719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.899806 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c"} err="failed to get container status \"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\": rpc error: code = NotFound desc = could not find container \"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\": container with ID starting with 719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.899828 4948 scope.go:117] "RemoveContainer" containerID="3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7" Feb 20 08:15:35 crc kubenswrapper[4948]: E0220 08:15:35.900123 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\": container with ID starting with 3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7 not found: ID does not exist" containerID="3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.900153 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7"} err="failed to get container status \"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\": rpc error: code = NotFound desc = could not find container \"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\": container with ID starting with 3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7 not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.900173 4948 scope.go:117] "RemoveContainer" containerID="9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d" Feb 20 08:15:35 crc kubenswrapper[4948]: E0220 08:15:35.900587 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\": container with ID starting with 9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d not found: ID does not exist" containerID="9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.900626 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d"} err="failed to get container status \"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\": rpc error: code = NotFound desc = could not find container \"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\": container with ID starting with 9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.900666 4948 scope.go:117] "RemoveContainer" containerID="540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad" Feb 20 08:15:35 crc kubenswrapper[4948]: E0220 08:15:35.900954 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\": container with ID starting with 540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad not found: ID does not exist" containerID="540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.901000 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad"} err="failed to get container status \"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\": rpc error: code = NotFound desc = could not find container \"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\": container with ID starting with 540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.901018 4948 scope.go:117] "RemoveContainer" containerID="682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.901305 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a"} err="failed to get container status \"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a\": rpc error: code = NotFound desc = could not find container \"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a\": container with ID starting with 682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.901331 4948 scope.go:117] "RemoveContainer" containerID="7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.901609 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b"} err="failed to get container status \"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b\": rpc error: code = NotFound desc = could not find container \"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b\": container with ID starting with 7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.901629 4948 scope.go:117] "RemoveContainer" containerID="c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.901866 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022"} err="failed to get container status \"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\": rpc error: code = NotFound desc = could not find container \"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\": container with ID starting with c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022 not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.901894 4948 scope.go:117] "RemoveContainer" containerID="8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.902198 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a"} err="failed to get container status \"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\": rpc error: code = NotFound desc = could not find container \"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\": container with ID starting with 8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.902225 4948 scope.go:117] "RemoveContainer" containerID="5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.902450 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2"} err="failed to get container status \"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\": rpc error: code = NotFound desc = could not find container \"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\": container with ID starting with 5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2 not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.902487 4948 scope.go:117] "RemoveContainer" containerID="6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.902685 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37"} err="failed to get container status \"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\": rpc error: code = NotFound desc = could not find container \"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\": container with ID starting with 6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37 not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.902717 4948 scope.go:117] "RemoveContainer" containerID="719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.902885 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c"} err="failed to get container status \"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\": rpc error: code = NotFound desc = could not find container \"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\": container with ID starting with 719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.902900 4948 scope.go:117] "RemoveContainer" containerID="3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.903208 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7"} err="failed to get container status \"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\": rpc error: code = NotFound desc = could not find container \"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\": container with ID starting with 3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7 not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.903287 4948 scope.go:117] "RemoveContainer" containerID="9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.903550 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d"} err="failed to get container status \"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\": rpc error: code = NotFound desc = could not find container \"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\": container with ID starting with 9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.903635 4948 scope.go:117] "RemoveContainer" containerID="540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.903884 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad"} err="failed to get container status \"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\": rpc error: code = NotFound desc = could not find container \"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\": container with ID starting with 540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.903965 4948 scope.go:117] "RemoveContainer" containerID="682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.904254 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a"} err="failed to get container status \"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a\": rpc error: code = NotFound desc = could not find container \"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a\": container with ID starting with 682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.904337 4948 scope.go:117] "RemoveContainer" containerID="7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.904648 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b"} err="failed to get container status \"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b\": rpc error: code = NotFound desc = could not find container \"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b\": container with ID starting with 7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.904678 4948 scope.go:117] "RemoveContainer" containerID="c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.904906 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022"} err="failed to get container status \"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\": rpc error: code = NotFound desc = could not find container \"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\": container with ID starting with c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022 not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.905006 4948 scope.go:117] "RemoveContainer" containerID="8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.905446 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a"} err="failed to get container status \"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\": rpc error: code = NotFound desc = could not find container \"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\": container with ID starting with 8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.905507 4948 scope.go:117] "RemoveContainer" containerID="5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.905856 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2"} err="failed to get container status \"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\": rpc error: code = NotFound desc = could not find container \"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\": container with ID starting with 5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2 not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.905881 4948 scope.go:117] "RemoveContainer" containerID="6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.906228 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37"} err="failed to get container status \"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\": rpc error: code = NotFound desc = could not find container \"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\": container with ID starting with 6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37 not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.906271 4948 scope.go:117] "RemoveContainer" containerID="719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.906579 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c"} err="failed to get container status \"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\": rpc error: code = NotFound desc = could not find container \"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\": container with ID starting with 719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.906605 4948 scope.go:117] "RemoveContainer" containerID="3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.906883 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7"} err="failed to get container status \"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\": rpc error: code = NotFound desc = could not find container \"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\": container with ID starting with 3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7 not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.906913 4948 scope.go:117] "RemoveContainer" containerID="9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.907213 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d"} err="failed to get container status \"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\": rpc error: code = NotFound desc = could not find container \"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\": container with ID starting with 9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.907239 4948 scope.go:117] "RemoveContainer" containerID="540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.907453 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad"} err="failed to get container status \"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\": rpc error: code = NotFound desc = could not find container \"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\": container with ID starting with 540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.907539 4948 scope.go:117] "RemoveContainer" containerID="682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.907770 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a"} err="failed to get container status \"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a\": rpc error: code = NotFound desc = could not find container \"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a\": container with ID starting with 682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.907795 4948 scope.go:117] "RemoveContainer" containerID="7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.908046 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b"} err="failed to get container status \"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b\": rpc error: code = NotFound desc = could not find container \"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b\": container with ID starting with 7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.908072 4948 scope.go:117] "RemoveContainer" containerID="c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.908248 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022"} err="failed to get container status \"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\": rpc error: code = NotFound desc = could not find container \"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\": container with ID starting with c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022 not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.908273 4948 scope.go:117] "RemoveContainer" containerID="8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.908452 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a"} err="failed to get container status \"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\": rpc error: code = NotFound desc = could not find container \"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\": container with ID starting with 8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.908475 4948 scope.go:117] "RemoveContainer" containerID="5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.908646 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2"} err="failed to get container status \"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\": rpc error: code = NotFound desc = could not find container \"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\": container with ID starting with 5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2 not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.908666 4948 scope.go:117] "RemoveContainer" containerID="6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.908846 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37"} err="failed to get container status \"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\": rpc error: code = NotFound desc = could not find container \"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\": container with ID starting with 6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37 not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.908871 4948 scope.go:117] "RemoveContainer" containerID="719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.909046 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c"} err="failed to get container status \"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\": rpc error: code = NotFound desc = could not find container \"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\": container with ID starting with 719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.909095 4948 scope.go:117] "RemoveContainer" containerID="3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.909252 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7"} err="failed to get container status \"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\": rpc error: code = NotFound desc = could not find container \"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\": container with ID starting with 3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7 not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.909276 4948 scope.go:117] "RemoveContainer" containerID="9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.909430 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d"} err="failed to get container status \"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\": rpc error: code = NotFound desc = could not find container \"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\": container with ID starting with 9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.909448 4948 scope.go:117] "RemoveContainer" containerID="540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.909606 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad"} err="failed to get container status \"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\": rpc error: code = NotFound desc = could not find container \"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\": container with ID starting with 540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.909636 4948 scope.go:117] "RemoveContainer" containerID="682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.909822 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a"} err="failed to get container status \"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a\": rpc error: code = NotFound desc = could not find container \"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a\": container with ID starting with 682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a not found: ID does not exist" Feb 20 08:15:36 crc kubenswrapper[4948]: I0220 08:15:36.625285 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" event={"ID":"0bfc877e-59ed-4803-b07a-b9787b1ab73d","Type":"ContainerStarted","Data":"fdde3da322ea47cd1001dc4dc1eb54b763197897461c6883cbe223d44b7b759a"} Feb 20 08:15:36 crc kubenswrapper[4948]: I0220 08:15:36.625639 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" event={"ID":"0bfc877e-59ed-4803-b07a-b9787b1ab73d","Type":"ContainerStarted","Data":"b41d6705ce7f06442f224bb66c7e09abe4a15233c1c95a1729006170bae9e8de"} Feb 20 08:15:36 crc kubenswrapper[4948]: I0220 08:15:36.625653 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" event={"ID":"0bfc877e-59ed-4803-b07a-b9787b1ab73d","Type":"ContainerStarted","Data":"a9d7900d2ba119eb35f9472059268560504ebcdb3f51d1d00c88119a2d0ceae9"} Feb 20 08:15:36 crc kubenswrapper[4948]: I0220 08:15:36.625663 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" event={"ID":"0bfc877e-59ed-4803-b07a-b9787b1ab73d","Type":"ContainerStarted","Data":"d8ad6bba55ee80255bca1c489ba329f108b4252e8eec315f332302ef06af7bcd"} Feb 20 08:15:36 crc kubenswrapper[4948]: I0220 08:15:36.625671 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" event={"ID":"0bfc877e-59ed-4803-b07a-b9787b1ab73d","Type":"ContainerStarted","Data":"f52c5394617a409495a9f9b19d57e6549b2a745ab62eaeaf69a3e4f2f9e34f3b"} Feb 20 08:15:36 crc kubenswrapper[4948]: I0220 08:15:36.625680 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" event={"ID":"0bfc877e-59ed-4803-b07a-b9787b1ab73d","Type":"ContainerStarted","Data":"bccf823b97fcab3fd9a064d741fe0510b30d622a5a022d73e528e79d103d68bc"} Feb 20 08:15:38 crc kubenswrapper[4948]: I0220 08:15:38.025207 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:15:38 crc kubenswrapper[4948]: I0220 08:15:38.025306 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:15:38 crc kubenswrapper[4948]: I0220 08:15:38.025377 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:15:38 crc kubenswrapper[4948]: I0220 08:15:38.026288 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a2c247070dfb4d0488751bb46e5543e834b3853c983d98f00e419c6aabfdeb7c"} pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 08:15:38 crc kubenswrapper[4948]: I0220 08:15:38.026394 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" containerID="cri-o://a2c247070dfb4d0488751bb46e5543e834b3853c983d98f00e419c6aabfdeb7c" gracePeriod=600 Feb 20 08:15:38 crc kubenswrapper[4948]: I0220 08:15:38.644699 4948 generic.go:334] "Generic (PLEG): container finished" podID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerID="a2c247070dfb4d0488751bb46e5543e834b3853c983d98f00e419c6aabfdeb7c" exitCode=0 Feb 20 08:15:38 crc kubenswrapper[4948]: I0220 08:15:38.644752 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerDied","Data":"a2c247070dfb4d0488751bb46e5543e834b3853c983d98f00e419c6aabfdeb7c"} Feb 20 08:15:38 crc kubenswrapper[4948]: I0220 08:15:38.645463 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"b5405a03f834a339ac758e6efb1e31a8c094b3f6b973ea8540eefc716ba7bf7f"} Feb 20 08:15:38 crc kubenswrapper[4948]: I0220 08:15:38.645501 4948 scope.go:117] "RemoveContainer" containerID="6791802a782456e3c45df951fa55097ce0ea3affd6a9878fc2981511046d1b09" Feb 20 08:15:39 crc kubenswrapper[4948]: I0220 08:15:39.659604 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" event={"ID":"0bfc877e-59ed-4803-b07a-b9787b1ab73d","Type":"ContainerStarted","Data":"15c07d823d4ea0c40991b497c78c8d77a94305232ecc7462264b9b70da7d7f0b"} Feb 20 08:15:41 crc kubenswrapper[4948]: I0220 08:15:41.678904 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" event={"ID":"0bfc877e-59ed-4803-b07a-b9787b1ab73d","Type":"ContainerStarted","Data":"e1ed772a07c900cbc9a9b6fa2cca2ea52073eff0653e8ed0b21f025559c1087e"} Feb 20 08:15:41 crc kubenswrapper[4948]: I0220 08:15:41.679517 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:41 crc kubenswrapper[4948]: I0220 08:15:41.720220 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" podStartSLOduration=7.720203128 podStartE2EDuration="7.720203128s" podCreationTimestamp="2026-02-20 08:15:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:15:41.714045724 +0000 UTC m=+590.688540574" watchObservedRunningTime="2026-02-20 08:15:41.720203128 +0000 UTC m=+590.694697948" Feb 20 08:15:41 crc kubenswrapper[4948]: I0220 08:15:41.747309 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:42 crc kubenswrapper[4948]: I0220 08:15:42.687887 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:42 crc kubenswrapper[4948]: I0220 08:15:42.688319 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:42 crc kubenswrapper[4948]: I0220 08:15:42.775082 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:47 crc kubenswrapper[4948]: I0220 08:15:47.723182 4948 scope.go:117] "RemoveContainer" containerID="4ecf8c7c9ed34410ff4619b18254e2eb90170911300e6cd6a7284e84e8937246" Feb 20 08:15:47 crc kubenswrapper[4948]: E0220 08:15:47.723669 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-x7s56_openshift-multus(b96124e4-0a74-4578-9142-fd728eb9f99e)\"" pod="openshift-multus/multus-x7s56" podUID="b96124e4-0a74-4578-9142-fd728eb9f99e" Feb 20 08:15:52 crc kubenswrapper[4948]: I0220 08:15:52.003509 4948 scope.go:117] "RemoveContainer" containerID="2f4093ec9d71181c58ad80a8c4ee3692148590df14aadaec96c0f54a843d1f8c" Feb 20 08:15:52 crc kubenswrapper[4948]: I0220 08:15:52.761470 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x7s56_b96124e4-0a74-4578-9142-fd728eb9f99e/kube-multus/2.log" Feb 20 08:15:58 crc kubenswrapper[4948]: I0220 08:15:58.723079 4948 scope.go:117] "RemoveContainer" containerID="4ecf8c7c9ed34410ff4619b18254e2eb90170911300e6cd6a7284e84e8937246" Feb 20 08:15:59 crc kubenswrapper[4948]: I0220 08:15:59.814731 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x7s56_b96124e4-0a74-4578-9142-fd728eb9f99e/kube-multus/2.log" Feb 20 08:15:59 crc kubenswrapper[4948]: I0220 08:15:59.815102 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x7s56" event={"ID":"b96124e4-0a74-4578-9142-fd728eb9f99e","Type":"ContainerStarted","Data":"e509f78ee7018faa8829b9687c5ed7446ec2bd36a0ef7fc02e9a2a538c8be56d"} Feb 20 08:16:05 crc kubenswrapper[4948]: I0220 08:16:05.307159 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:16:12 crc kubenswrapper[4948]: I0220 08:16:12.077284 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w"] Feb 20 08:16:12 crc kubenswrapper[4948]: I0220 08:16:12.078925 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" Feb 20 08:16:12 crc kubenswrapper[4948]: I0220 08:16:12.082225 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 20 08:16:12 crc kubenswrapper[4948]: I0220 08:16:12.087995 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w"] Feb 20 08:16:12 crc kubenswrapper[4948]: I0220 08:16:12.204284 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvwb7\" (UniqueName: \"kubernetes.io/projected/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-kube-api-access-nvwb7\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w\" (UID: \"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" Feb 20 08:16:12 crc kubenswrapper[4948]: I0220 08:16:12.204381 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-bundle\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w\" (UID: \"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" Feb 20 08:16:12 crc kubenswrapper[4948]: I0220 08:16:12.204412 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-util\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w\" (UID: \"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" Feb 20 08:16:12 crc kubenswrapper[4948]: I0220 08:16:12.305491 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-bundle\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w\" (UID: \"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" Feb 20 08:16:12 crc kubenswrapper[4948]: I0220 08:16:12.305564 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-util\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w\" (UID: \"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" Feb 20 08:16:12 crc kubenswrapper[4948]: I0220 08:16:12.305684 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvwb7\" (UniqueName: \"kubernetes.io/projected/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-kube-api-access-nvwb7\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w\" (UID: \"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" Feb 20 08:16:12 crc kubenswrapper[4948]: I0220 08:16:12.306550 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-bundle\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w\" (UID: \"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" Feb 20 08:16:12 crc kubenswrapper[4948]: I0220 08:16:12.306741 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-util\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w\" (UID: \"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" Feb 20 08:16:12 crc kubenswrapper[4948]: I0220 08:16:12.338270 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvwb7\" (UniqueName: \"kubernetes.io/projected/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-kube-api-access-nvwb7\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w\" (UID: \"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" Feb 20 08:16:12 crc kubenswrapper[4948]: I0220 08:16:12.398273 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" Feb 20 08:16:12 crc kubenswrapper[4948]: I0220 08:16:12.673937 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w"] Feb 20 08:16:12 crc kubenswrapper[4948]: W0220 08:16:12.681901 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb3ec19c_541d_4b32_9f5c_ac61f28bdc2a.slice/crio-548722355d09a176b724709bdf45edc717f09882503ef4ddac12a3f79190d30d WatchSource:0}: Error finding container 548722355d09a176b724709bdf45edc717f09882503ef4ddac12a3f79190d30d: Status 404 returned error can't find the container with id 548722355d09a176b724709bdf45edc717f09882503ef4ddac12a3f79190d30d Feb 20 08:16:12 crc kubenswrapper[4948]: I0220 08:16:12.898185 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" event={"ID":"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a","Type":"ContainerStarted","Data":"23679add441ce28e94c0d0fe925e7ecfaf8abe0638184164f4d2902ffb5cfa86"} Feb 20 08:16:12 crc kubenswrapper[4948]: I0220 08:16:12.898253 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" event={"ID":"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a","Type":"ContainerStarted","Data":"548722355d09a176b724709bdf45edc717f09882503ef4ddac12a3f79190d30d"} Feb 20 08:16:13 crc kubenswrapper[4948]: I0220 08:16:13.905260 4948 generic.go:334] "Generic (PLEG): container finished" podID="bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a" containerID="23679add441ce28e94c0d0fe925e7ecfaf8abe0638184164f4d2902ffb5cfa86" exitCode=0 Feb 20 08:16:13 crc kubenswrapper[4948]: I0220 08:16:13.905317 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" event={"ID":"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a","Type":"ContainerDied","Data":"23679add441ce28e94c0d0fe925e7ecfaf8abe0638184164f4d2902ffb5cfa86"} Feb 20 08:16:15 crc kubenswrapper[4948]: I0220 08:16:15.922364 4948 generic.go:334] "Generic (PLEG): container finished" podID="bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a" containerID="4b79219701d24d57bef90a1d915b063c7b1bfa1fad99516543195699165a1cc8" exitCode=0 Feb 20 08:16:15 crc kubenswrapper[4948]: I0220 08:16:15.922424 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" event={"ID":"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a","Type":"ContainerDied","Data":"4b79219701d24d57bef90a1d915b063c7b1bfa1fad99516543195699165a1cc8"} Feb 20 08:16:16 crc kubenswrapper[4948]: I0220 08:16:16.932125 4948 generic.go:334] "Generic (PLEG): container finished" podID="bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a" containerID="8dd2cd73dd9334db76086a27f76ad986432450f60bbc24bf1cebe547d08c3e88" exitCode=0 Feb 20 08:16:16 crc kubenswrapper[4948]: I0220 08:16:16.932170 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" event={"ID":"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a","Type":"ContainerDied","Data":"8dd2cd73dd9334db76086a27f76ad986432450f60bbc24bf1cebe547d08c3e88"} Feb 20 08:16:18 crc kubenswrapper[4948]: I0220 08:16:18.217368 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" Feb 20 08:16:18 crc kubenswrapper[4948]: I0220 08:16:18.391781 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-util\") pod \"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a\" (UID: \"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a\") " Feb 20 08:16:18 crc kubenswrapper[4948]: I0220 08:16:18.391880 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvwb7\" (UniqueName: \"kubernetes.io/projected/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-kube-api-access-nvwb7\") pod \"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a\" (UID: \"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a\") " Feb 20 08:16:18 crc kubenswrapper[4948]: I0220 08:16:18.391955 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-bundle\") pod \"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a\" (UID: \"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a\") " Feb 20 08:16:18 crc kubenswrapper[4948]: I0220 08:16:18.393130 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-bundle" (OuterVolumeSpecName: "bundle") pod "bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a" (UID: "bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:16:18 crc kubenswrapper[4948]: I0220 08:16:18.398721 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-kube-api-access-nvwb7" (OuterVolumeSpecName: "kube-api-access-nvwb7") pod "bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a" (UID: "bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a"). InnerVolumeSpecName "kube-api-access-nvwb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:16:18 crc kubenswrapper[4948]: I0220 08:16:18.476391 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-util" (OuterVolumeSpecName: "util") pod "bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a" (UID: "bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:16:18 crc kubenswrapper[4948]: I0220 08:16:18.493722 4948 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-util\") on node \"crc\" DevicePath \"\"" Feb 20 08:16:18 crc kubenswrapper[4948]: I0220 08:16:18.493759 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvwb7\" (UniqueName: \"kubernetes.io/projected/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-kube-api-access-nvwb7\") on node \"crc\" DevicePath \"\"" Feb 20 08:16:18 crc kubenswrapper[4948]: I0220 08:16:18.493780 4948 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:16:18 crc kubenswrapper[4948]: I0220 08:16:18.948617 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" event={"ID":"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a","Type":"ContainerDied","Data":"548722355d09a176b724709bdf45edc717f09882503ef4ddac12a3f79190d30d"} Feb 20 08:16:18 crc kubenswrapper[4948]: I0220 08:16:18.948675 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="548722355d09a176b724709bdf45edc717f09882503ef4ddac12a3f79190d30d" Feb 20 08:16:18 crc kubenswrapper[4948]: I0220 08:16:18.948783 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" Feb 20 08:16:21 crc kubenswrapper[4948]: I0220 08:16:21.023253 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-694c9596b7-xqtvg"] Feb 20 08:16:21 crc kubenswrapper[4948]: E0220 08:16:21.023896 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a" containerName="util" Feb 20 08:16:21 crc kubenswrapper[4948]: I0220 08:16:21.023910 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a" containerName="util" Feb 20 08:16:21 crc kubenswrapper[4948]: E0220 08:16:21.023925 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a" containerName="pull" Feb 20 08:16:21 crc kubenswrapper[4948]: I0220 08:16:21.023933 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a" containerName="pull" Feb 20 08:16:21 crc kubenswrapper[4948]: E0220 08:16:21.023943 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a" containerName="extract" Feb 20 08:16:21 crc kubenswrapper[4948]: I0220 08:16:21.023952 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a" containerName="extract" Feb 20 08:16:21 crc kubenswrapper[4948]: I0220 08:16:21.024112 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a" containerName="extract" Feb 20 08:16:21 crc kubenswrapper[4948]: I0220 08:16:21.024535 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-694c9596b7-xqtvg" Feb 20 08:16:21 crc kubenswrapper[4948]: I0220 08:16:21.026079 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w25ws\" (UniqueName: \"kubernetes.io/projected/59277281-0112-45da-a64a-1b1d02b90473-kube-api-access-w25ws\") pod \"nmstate-operator-694c9596b7-xqtvg\" (UID: \"59277281-0112-45da-a64a-1b1d02b90473\") " pod="openshift-nmstate/nmstate-operator-694c9596b7-xqtvg" Feb 20 08:16:21 crc kubenswrapper[4948]: I0220 08:16:21.027247 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-vkzsw" Feb 20 08:16:21 crc kubenswrapper[4948]: I0220 08:16:21.027673 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Feb 20 08:16:21 crc kubenswrapper[4948]: I0220 08:16:21.028449 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Feb 20 08:16:21 crc kubenswrapper[4948]: I0220 08:16:21.036738 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-694c9596b7-xqtvg"] Feb 20 08:16:21 crc kubenswrapper[4948]: I0220 08:16:21.127445 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w25ws\" (UniqueName: \"kubernetes.io/projected/59277281-0112-45da-a64a-1b1d02b90473-kube-api-access-w25ws\") pod \"nmstate-operator-694c9596b7-xqtvg\" (UID: \"59277281-0112-45da-a64a-1b1d02b90473\") " pod="openshift-nmstate/nmstate-operator-694c9596b7-xqtvg" Feb 20 08:16:21 crc kubenswrapper[4948]: I0220 08:16:21.151742 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w25ws\" (UniqueName: \"kubernetes.io/projected/59277281-0112-45da-a64a-1b1d02b90473-kube-api-access-w25ws\") pod \"nmstate-operator-694c9596b7-xqtvg\" (UID: \"59277281-0112-45da-a64a-1b1d02b90473\") " pod="openshift-nmstate/nmstate-operator-694c9596b7-xqtvg" Feb 20 08:16:21 crc kubenswrapper[4948]: I0220 08:16:21.350483 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-694c9596b7-xqtvg" Feb 20 08:16:21 crc kubenswrapper[4948]: I0220 08:16:21.529077 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-694c9596b7-xqtvg"] Feb 20 08:16:21 crc kubenswrapper[4948]: I0220 08:16:21.966606 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-694c9596b7-xqtvg" event={"ID":"59277281-0112-45da-a64a-1b1d02b90473","Type":"ContainerStarted","Data":"c05b6a8309c306ed7cfbbb7e06e1c7f43f95c3a41836739a15a14819c2700731"} Feb 20 08:16:28 crc kubenswrapper[4948]: I0220 08:16:28.014849 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-694c9596b7-xqtvg" event={"ID":"59277281-0112-45da-a64a-1b1d02b90473","Type":"ContainerStarted","Data":"4467fa79cf451a4a49018633dda15922cfc832112e00e679a559ad7e775fb518"} Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.059138 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-694c9596b7-xqtvg" podStartSLOduration=3.163767336 podStartE2EDuration="9.059112218s" podCreationTimestamp="2026-02-20 08:16:20 +0000 UTC" firstStartedPulling="2026-02-20 08:16:21.534283725 +0000 UTC m=+630.508778545" lastFinishedPulling="2026-02-20 08:16:27.429628597 +0000 UTC m=+636.404123427" observedRunningTime="2026-02-20 08:16:28.046165941 +0000 UTC m=+637.020660821" watchObservedRunningTime="2026-02-20 08:16:29.059112218 +0000 UTC m=+638.033607068" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.063696 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-58c85c668d-tdst7"] Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.065120 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58c85c668d-tdst7" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.067263 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-xqmp6" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.097052 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-dxdh6"] Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.098626 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-dxdh6" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.102211 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58c85c668d-tdst7"] Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.113619 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-866bcb46dc-77sdv"] Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.114775 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-77sdv" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.126811 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.160775 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-866bcb46dc-77sdv"] Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.212243 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5c78fc5d65-9mv9z"] Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.212902 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-9mv9z" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.215919 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.216207 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.222108 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-llbnp" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.224078 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5c78fc5d65-9mv9z"] Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.248727 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/3a869b3a-b757-49ba-9096-2e562d980aae-dbus-socket\") pod \"nmstate-handler-dxdh6\" (UID: \"3a869b3a-b757-49ba-9096-2e562d980aae\") " pod="openshift-nmstate/nmstate-handler-dxdh6" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.248791 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/3a869b3a-b757-49ba-9096-2e562d980aae-ovs-socket\") pod \"nmstate-handler-dxdh6\" (UID: \"3a869b3a-b757-49ba-9096-2e562d980aae\") " pod="openshift-nmstate/nmstate-handler-dxdh6" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.248841 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/da0e4284-f20d-44a8-b529-e388f15822d3-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-77sdv\" (UID: \"da0e4284-f20d-44a8-b529-e388f15822d3\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-77sdv" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.248873 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67xdv\" (UniqueName: \"kubernetes.io/projected/3a869b3a-b757-49ba-9096-2e562d980aae-kube-api-access-67xdv\") pod \"nmstate-handler-dxdh6\" (UID: \"3a869b3a-b757-49ba-9096-2e562d980aae\") " pod="openshift-nmstate/nmstate-handler-dxdh6" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.248900 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/3a869b3a-b757-49ba-9096-2e562d980aae-nmstate-lock\") pod \"nmstate-handler-dxdh6\" (UID: \"3a869b3a-b757-49ba-9096-2e562d980aae\") " pod="openshift-nmstate/nmstate-handler-dxdh6" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.248922 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pp789\" (UniqueName: \"kubernetes.io/projected/da0e4284-f20d-44a8-b529-e388f15822d3-kube-api-access-pp789\") pod \"nmstate-webhook-866bcb46dc-77sdv\" (UID: \"da0e4284-f20d-44a8-b529-e388f15822d3\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-77sdv" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.248943 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tgvg\" (UniqueName: \"kubernetes.io/projected/a7b34c06-d0c5-4f20-9521-eec500ea2510-kube-api-access-6tgvg\") pod \"nmstate-metrics-58c85c668d-tdst7\" (UID: \"a7b34c06-d0c5-4f20-9521-eec500ea2510\") " pod="openshift-nmstate/nmstate-metrics-58c85c668d-tdst7" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.350559 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67xdv\" (UniqueName: \"kubernetes.io/projected/3a869b3a-b757-49ba-9096-2e562d980aae-kube-api-access-67xdv\") pod \"nmstate-handler-dxdh6\" (UID: \"3a869b3a-b757-49ba-9096-2e562d980aae\") " pod="openshift-nmstate/nmstate-handler-dxdh6" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.350629 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqt4s\" (UniqueName: \"kubernetes.io/projected/47461ed8-c51b-4014-bcf0-4d95a0278f85-kube-api-access-jqt4s\") pod \"nmstate-console-plugin-5c78fc5d65-9mv9z\" (UID: \"47461ed8-c51b-4014-bcf0-4d95a0278f85\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-9mv9z" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.350661 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/3a869b3a-b757-49ba-9096-2e562d980aae-nmstate-lock\") pod \"nmstate-handler-dxdh6\" (UID: \"3a869b3a-b757-49ba-9096-2e562d980aae\") " pod="openshift-nmstate/nmstate-handler-dxdh6" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.350695 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pp789\" (UniqueName: \"kubernetes.io/projected/da0e4284-f20d-44a8-b529-e388f15822d3-kube-api-access-pp789\") pod \"nmstate-webhook-866bcb46dc-77sdv\" (UID: \"da0e4284-f20d-44a8-b529-e388f15822d3\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-77sdv" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.350726 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tgvg\" (UniqueName: \"kubernetes.io/projected/a7b34c06-d0c5-4f20-9521-eec500ea2510-kube-api-access-6tgvg\") pod \"nmstate-metrics-58c85c668d-tdst7\" (UID: \"a7b34c06-d0c5-4f20-9521-eec500ea2510\") " pod="openshift-nmstate/nmstate-metrics-58c85c668d-tdst7" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.350778 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/3a869b3a-b757-49ba-9096-2e562d980aae-nmstate-lock\") pod \"nmstate-handler-dxdh6\" (UID: \"3a869b3a-b757-49ba-9096-2e562d980aae\") " pod="openshift-nmstate/nmstate-handler-dxdh6" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.350790 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/47461ed8-c51b-4014-bcf0-4d95a0278f85-nginx-conf\") pod \"nmstate-console-plugin-5c78fc5d65-9mv9z\" (UID: \"47461ed8-c51b-4014-bcf0-4d95a0278f85\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-9mv9z" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.350945 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/3a869b3a-b757-49ba-9096-2e562d980aae-dbus-socket\") pod \"nmstate-handler-dxdh6\" (UID: \"3a869b3a-b757-49ba-9096-2e562d980aae\") " pod="openshift-nmstate/nmstate-handler-dxdh6" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.351023 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/3a869b3a-b757-49ba-9096-2e562d980aae-ovs-socket\") pod \"nmstate-handler-dxdh6\" (UID: \"3a869b3a-b757-49ba-9096-2e562d980aae\") " pod="openshift-nmstate/nmstate-handler-dxdh6" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.351066 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/47461ed8-c51b-4014-bcf0-4d95a0278f85-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-9mv9z\" (UID: \"47461ed8-c51b-4014-bcf0-4d95a0278f85\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-9mv9z" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.351097 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/da0e4284-f20d-44a8-b529-e388f15822d3-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-77sdv\" (UID: \"da0e4284-f20d-44a8-b529-e388f15822d3\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-77sdv" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.351126 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/3a869b3a-b757-49ba-9096-2e562d980aae-ovs-socket\") pod \"nmstate-handler-dxdh6\" (UID: \"3a869b3a-b757-49ba-9096-2e562d980aae\") " pod="openshift-nmstate/nmstate-handler-dxdh6" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.351241 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/3a869b3a-b757-49ba-9096-2e562d980aae-dbus-socket\") pod \"nmstate-handler-dxdh6\" (UID: \"3a869b3a-b757-49ba-9096-2e562d980aae\") " pod="openshift-nmstate/nmstate-handler-dxdh6" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.361135 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/da0e4284-f20d-44a8-b529-e388f15822d3-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-77sdv\" (UID: \"da0e4284-f20d-44a8-b529-e388f15822d3\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-77sdv" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.371190 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67xdv\" (UniqueName: \"kubernetes.io/projected/3a869b3a-b757-49ba-9096-2e562d980aae-kube-api-access-67xdv\") pod \"nmstate-handler-dxdh6\" (UID: \"3a869b3a-b757-49ba-9096-2e562d980aae\") " pod="openshift-nmstate/nmstate-handler-dxdh6" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.375008 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tgvg\" (UniqueName: \"kubernetes.io/projected/a7b34c06-d0c5-4f20-9521-eec500ea2510-kube-api-access-6tgvg\") pod \"nmstate-metrics-58c85c668d-tdst7\" (UID: \"a7b34c06-d0c5-4f20-9521-eec500ea2510\") " pod="openshift-nmstate/nmstate-metrics-58c85c668d-tdst7" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.381563 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pp789\" (UniqueName: \"kubernetes.io/projected/da0e4284-f20d-44a8-b529-e388f15822d3-kube-api-access-pp789\") pod \"nmstate-webhook-866bcb46dc-77sdv\" (UID: \"da0e4284-f20d-44a8-b529-e388f15822d3\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-77sdv" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.405336 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58c85c668d-tdst7" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.428718 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-dxdh6" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.446026 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-77sdv" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.449323 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-64dcb9c7f-fwdtz"] Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.450390 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.452118 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/47461ed8-c51b-4014-bcf0-4d95a0278f85-nginx-conf\") pod \"nmstate-console-plugin-5c78fc5d65-9mv9z\" (UID: \"47461ed8-c51b-4014-bcf0-4d95a0278f85\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-9mv9z" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.452187 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/47461ed8-c51b-4014-bcf0-4d95a0278f85-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-9mv9z\" (UID: \"47461ed8-c51b-4014-bcf0-4d95a0278f85\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-9mv9z" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.452234 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqt4s\" (UniqueName: \"kubernetes.io/projected/47461ed8-c51b-4014-bcf0-4d95a0278f85-kube-api-access-jqt4s\") pod \"nmstate-console-plugin-5c78fc5d65-9mv9z\" (UID: \"47461ed8-c51b-4014-bcf0-4d95a0278f85\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-9mv9z" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.453470 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/47461ed8-c51b-4014-bcf0-4d95a0278f85-nginx-conf\") pod \"nmstate-console-plugin-5c78fc5d65-9mv9z\" (UID: \"47461ed8-c51b-4014-bcf0-4d95a0278f85\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-9mv9z" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.481196 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqt4s\" (UniqueName: \"kubernetes.io/projected/47461ed8-c51b-4014-bcf0-4d95a0278f85-kube-api-access-jqt4s\") pod \"nmstate-console-plugin-5c78fc5d65-9mv9z\" (UID: \"47461ed8-c51b-4014-bcf0-4d95a0278f85\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-9mv9z" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.481219 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/47461ed8-c51b-4014-bcf0-4d95a0278f85-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-9mv9z\" (UID: \"47461ed8-c51b-4014-bcf0-4d95a0278f85\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-9mv9z" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.486852 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-64dcb9c7f-fwdtz"] Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.533756 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-9mv9z" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.553420 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9b67f523-c74e-4b03-99cf-64fe0dd981f2-console-config\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.553477 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhtt6\" (UniqueName: \"kubernetes.io/projected/9b67f523-c74e-4b03-99cf-64fe0dd981f2-kube-api-access-mhtt6\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.553575 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9b67f523-c74e-4b03-99cf-64fe0dd981f2-trusted-ca-bundle\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.553648 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9b67f523-c74e-4b03-99cf-64fe0dd981f2-oauth-serving-cert\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.553684 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9b67f523-c74e-4b03-99cf-64fe0dd981f2-service-ca\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.553933 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9b67f523-c74e-4b03-99cf-64fe0dd981f2-console-serving-cert\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.554036 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9b67f523-c74e-4b03-99cf-64fe0dd981f2-console-oauth-config\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.639508 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58c85c668d-tdst7"] Feb 20 08:16:29 crc kubenswrapper[4948]: W0220 08:16:29.653229 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda7b34c06_d0c5_4f20_9521_eec500ea2510.slice/crio-2f69a3cdfed473617a7a51a1fd6975716f998a88d66b252db4a98122759defcd WatchSource:0}: Error finding container 2f69a3cdfed473617a7a51a1fd6975716f998a88d66b252db4a98122759defcd: Status 404 returned error can't find the container with id 2f69a3cdfed473617a7a51a1fd6975716f998a88d66b252db4a98122759defcd Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.654805 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhtt6\" (UniqueName: \"kubernetes.io/projected/9b67f523-c74e-4b03-99cf-64fe0dd981f2-kube-api-access-mhtt6\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.654848 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9b67f523-c74e-4b03-99cf-64fe0dd981f2-trusted-ca-bundle\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.654876 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9b67f523-c74e-4b03-99cf-64fe0dd981f2-oauth-serving-cert\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.654895 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9b67f523-c74e-4b03-99cf-64fe0dd981f2-service-ca\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.654939 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9b67f523-c74e-4b03-99cf-64fe0dd981f2-console-serving-cert\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.654982 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9b67f523-c74e-4b03-99cf-64fe0dd981f2-console-oauth-config\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.655002 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9b67f523-c74e-4b03-99cf-64fe0dd981f2-console-config\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.656575 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9b67f523-c74e-4b03-99cf-64fe0dd981f2-oauth-serving-cert\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.656966 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9b67f523-c74e-4b03-99cf-64fe0dd981f2-trusted-ca-bundle\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.657070 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9b67f523-c74e-4b03-99cf-64fe0dd981f2-console-config\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.657733 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9b67f523-c74e-4b03-99cf-64fe0dd981f2-service-ca\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.659873 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9b67f523-c74e-4b03-99cf-64fe0dd981f2-console-serving-cert\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.669586 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9b67f523-c74e-4b03-99cf-64fe0dd981f2-console-oauth-config\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.673944 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhtt6\" (UniqueName: \"kubernetes.io/projected/9b67f523-c74e-4b03-99cf-64fe0dd981f2-kube-api-access-mhtt6\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: W0220 08:16:29.781794 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod47461ed8_c51b_4014_bcf0_4d95a0278f85.slice/crio-7d31361835d89022a27dbe1075d24c03db175adbb0decc2479704bd81951c240 WatchSource:0}: Error finding container 7d31361835d89022a27dbe1075d24c03db175adbb0decc2479704bd81951c240: Status 404 returned error can't find the container with id 7d31361835d89022a27dbe1075d24c03db175adbb0decc2479704bd81951c240 Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.782771 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5c78fc5d65-9mv9z"] Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.804516 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.950793 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-866bcb46dc-77sdv"] Feb 20 08:16:29 crc kubenswrapper[4948]: W0220 08:16:29.959403 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podda0e4284_f20d_44a8_b529_e388f15822d3.slice/crio-b8fa1601d3f454168f034271bfc5dc2d68a4cc6611e8568c6cff0f11fb9e443d WatchSource:0}: Error finding container b8fa1601d3f454168f034271bfc5dc2d68a4cc6611e8568c6cff0f11fb9e443d: Status 404 returned error can't find the container with id b8fa1601d3f454168f034271bfc5dc2d68a4cc6611e8568c6cff0f11fb9e443d Feb 20 08:16:30 crc kubenswrapper[4948]: I0220 08:16:30.029106 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-9mv9z" event={"ID":"47461ed8-c51b-4014-bcf0-4d95a0278f85","Type":"ContainerStarted","Data":"7d31361835d89022a27dbe1075d24c03db175adbb0decc2479704bd81951c240"} Feb 20 08:16:30 crc kubenswrapper[4948]: I0220 08:16:30.030708 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58c85c668d-tdst7" event={"ID":"a7b34c06-d0c5-4f20-9521-eec500ea2510","Type":"ContainerStarted","Data":"2f69a3cdfed473617a7a51a1fd6975716f998a88d66b252db4a98122759defcd"} Feb 20 08:16:30 crc kubenswrapper[4948]: I0220 08:16:30.031788 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-77sdv" event={"ID":"da0e4284-f20d-44a8-b529-e388f15822d3","Type":"ContainerStarted","Data":"b8fa1601d3f454168f034271bfc5dc2d68a4cc6611e8568c6cff0f11fb9e443d"} Feb 20 08:16:30 crc kubenswrapper[4948]: I0220 08:16:30.033716 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-dxdh6" event={"ID":"3a869b3a-b757-49ba-9096-2e562d980aae","Type":"ContainerStarted","Data":"3c9a321ad1b783cb64857b52dd5a94a4cc043894390bb304392d324d8553e8fc"} Feb 20 08:16:30 crc kubenswrapper[4948]: I0220 08:16:30.058283 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-64dcb9c7f-fwdtz"] Feb 20 08:16:30 crc kubenswrapper[4948]: W0220 08:16:30.063130 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b67f523_c74e_4b03_99cf_64fe0dd981f2.slice/crio-b7245036c705e569134bea807738097ccff83132fb89b9e2861c36fcfe1ef775 WatchSource:0}: Error finding container b7245036c705e569134bea807738097ccff83132fb89b9e2861c36fcfe1ef775: Status 404 returned error can't find the container with id b7245036c705e569134bea807738097ccff83132fb89b9e2861c36fcfe1ef775 Feb 20 08:16:31 crc kubenswrapper[4948]: I0220 08:16:31.042139 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-64dcb9c7f-fwdtz" event={"ID":"9b67f523-c74e-4b03-99cf-64fe0dd981f2","Type":"ContainerStarted","Data":"5b4b88a4af63b9a4d22a23d19cc8005963d21d10edd08aa88ee963fbb6f2a798"} Feb 20 08:16:31 crc kubenswrapper[4948]: I0220 08:16:31.042440 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-64dcb9c7f-fwdtz" event={"ID":"9b67f523-c74e-4b03-99cf-64fe0dd981f2","Type":"ContainerStarted","Data":"b7245036c705e569134bea807738097ccff83132fb89b9e2861c36fcfe1ef775"} Feb 20 08:16:31 crc kubenswrapper[4948]: I0220 08:16:31.756591 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-64dcb9c7f-fwdtz" podStartSLOduration=2.756564242 podStartE2EDuration="2.756564242s" podCreationTimestamp="2026-02-20 08:16:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:16:31.06452074 +0000 UTC m=+640.039015560" watchObservedRunningTime="2026-02-20 08:16:31.756564242 +0000 UTC m=+640.731059102" Feb 20 08:16:33 crc kubenswrapper[4948]: I0220 08:16:33.059049 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-9mv9z" event={"ID":"47461ed8-c51b-4014-bcf0-4d95a0278f85","Type":"ContainerStarted","Data":"d126b06cf7663226603898b9659ec53aa29323f2f8277118fa53815a5cd83895"} Feb 20 08:16:33 crc kubenswrapper[4948]: I0220 08:16:33.070930 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58c85c668d-tdst7" event={"ID":"a7b34c06-d0c5-4f20-9521-eec500ea2510","Type":"ContainerStarted","Data":"690f316803c2ecde318a6e40f63ad3ba7cd1be8c5b0541c5f60be03c437f23d3"} Feb 20 08:16:33 crc kubenswrapper[4948]: I0220 08:16:33.073499 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-77sdv" event={"ID":"da0e4284-f20d-44a8-b529-e388f15822d3","Type":"ContainerStarted","Data":"30d1210c199eb6e76765b57e55c646439da7610a9659e99f3f03629f2824d553"} Feb 20 08:16:33 crc kubenswrapper[4948]: I0220 08:16:33.073896 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-77sdv" Feb 20 08:16:33 crc kubenswrapper[4948]: I0220 08:16:33.077387 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-dxdh6" event={"ID":"3a869b3a-b757-49ba-9096-2e562d980aae","Type":"ContainerStarted","Data":"6e3aa8eb0ab95c174b74ca9e6b49ff3ed3735f50ca9cd7ba5ed209d3e8bfcf59"} Feb 20 08:16:33 crc kubenswrapper[4948]: I0220 08:16:33.078046 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-dxdh6" Feb 20 08:16:33 crc kubenswrapper[4948]: I0220 08:16:33.085361 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-9mv9z" podStartSLOduration=1.5106040200000002 podStartE2EDuration="4.085344586s" podCreationTimestamp="2026-02-20 08:16:29 +0000 UTC" firstStartedPulling="2026-02-20 08:16:29.784347201 +0000 UTC m=+638.758842011" lastFinishedPulling="2026-02-20 08:16:32.359087757 +0000 UTC m=+641.333582577" observedRunningTime="2026-02-20 08:16:33.084569997 +0000 UTC m=+642.059064827" watchObservedRunningTime="2026-02-20 08:16:33.085344586 +0000 UTC m=+642.059839446" Feb 20 08:16:33 crc kubenswrapper[4948]: I0220 08:16:33.116356 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-77sdv" podStartSLOduration=1.729020261 podStartE2EDuration="4.116331261s" podCreationTimestamp="2026-02-20 08:16:29 +0000 UTC" firstStartedPulling="2026-02-20 08:16:29.962638169 +0000 UTC m=+638.937132999" lastFinishedPulling="2026-02-20 08:16:32.349949139 +0000 UTC m=+641.324443999" observedRunningTime="2026-02-20 08:16:33.111405518 +0000 UTC m=+642.085900368" watchObservedRunningTime="2026-02-20 08:16:33.116331261 +0000 UTC m=+642.090826091" Feb 20 08:16:33 crc kubenswrapper[4948]: I0220 08:16:33.135387 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-dxdh6" podStartSLOduration=1.277670536 podStartE2EDuration="4.135368527s" podCreationTimestamp="2026-02-20 08:16:29 +0000 UTC" firstStartedPulling="2026-02-20 08:16:29.504666378 +0000 UTC m=+638.479161198" lastFinishedPulling="2026-02-20 08:16:32.362364369 +0000 UTC m=+641.336859189" observedRunningTime="2026-02-20 08:16:33.128027203 +0000 UTC m=+642.102522063" watchObservedRunningTime="2026-02-20 08:16:33.135368527 +0000 UTC m=+642.109863367" Feb 20 08:16:35 crc kubenswrapper[4948]: I0220 08:16:35.094569 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58c85c668d-tdst7" event={"ID":"a7b34c06-d0c5-4f20-9521-eec500ea2510","Type":"ContainerStarted","Data":"fd7928ff068e20efed3dba24766aca8d8b2797500a32b05fdb406cc49dcc9cab"} Feb 20 08:16:35 crc kubenswrapper[4948]: I0220 08:16:35.119804 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-58c85c668d-tdst7" podStartSLOduration=1.2550912300000001 podStartE2EDuration="6.119788593s" podCreationTimestamp="2026-02-20 08:16:29 +0000 UTC" firstStartedPulling="2026-02-20 08:16:29.657638533 +0000 UTC m=+638.632133353" lastFinishedPulling="2026-02-20 08:16:34.522335896 +0000 UTC m=+643.496830716" observedRunningTime="2026-02-20 08:16:35.11568264 +0000 UTC m=+644.090177460" watchObservedRunningTime="2026-02-20 08:16:35.119788593 +0000 UTC m=+644.094283413" Feb 20 08:16:39 crc kubenswrapper[4948]: I0220 08:16:39.468639 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-dxdh6" Feb 20 08:16:39 crc kubenswrapper[4948]: I0220 08:16:39.804644 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:39 crc kubenswrapper[4948]: I0220 08:16:39.804725 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:39 crc kubenswrapper[4948]: I0220 08:16:39.810604 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:40 crc kubenswrapper[4948]: I0220 08:16:40.134277 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:40 crc kubenswrapper[4948]: I0220 08:16:40.194462 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-k8zcr"] Feb 20 08:16:49 crc kubenswrapper[4948]: I0220 08:16:49.457172 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-77sdv" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.263137 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-k8zcr" podUID="f81366e5-fbdb-4a0b-a91f-f603d9b8570d" containerName="console" containerID="cri-o://3745ecb9e997c9c1a842fb71042af1df4c9828447494e128f09e0903f609bd78" gracePeriod=15 Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.597144 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-k8zcr_f81366e5-fbdb-4a0b-a91f-f603d9b8570d/console/0.log" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.597493 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.627421 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct"] Feb 20 08:17:05 crc kubenswrapper[4948]: E0220 08:17:05.627656 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f81366e5-fbdb-4a0b-a91f-f603d9b8570d" containerName="console" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.627667 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f81366e5-fbdb-4a0b-a91f-f603d9b8570d" containerName="console" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.627768 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f81366e5-fbdb-4a0b-a91f-f603d9b8570d" containerName="console" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.628445 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.630408 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.637416 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct"] Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.789351 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-oauth-serving-cert\") pod \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.789500 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7b5rg\" (UniqueName: \"kubernetes.io/projected/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-kube-api-access-7b5rg\") pod \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.789542 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-trusted-ca-bundle\") pod \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.789580 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-config\") pod \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.789669 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-service-ca\") pod \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.789718 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-serving-cert\") pod \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.789796 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-oauth-config\") pod \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.790063 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct\" (UID: \"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.790132 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcj8g\" (UniqueName: \"kubernetes.io/projected/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-kube-api-access-lcj8g\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct\" (UID: \"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.790217 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct\" (UID: \"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.790826 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-config" (OuterVolumeSpecName: "console-config") pod "f81366e5-fbdb-4a0b-a91f-f603d9b8570d" (UID: "f81366e5-fbdb-4a0b-a91f-f603d9b8570d"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.791014 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-service-ca" (OuterVolumeSpecName: "service-ca") pod "f81366e5-fbdb-4a0b-a91f-f603d9b8570d" (UID: "f81366e5-fbdb-4a0b-a91f-f603d9b8570d"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.791637 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "f81366e5-fbdb-4a0b-a91f-f603d9b8570d" (UID: "f81366e5-fbdb-4a0b-a91f-f603d9b8570d"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.791690 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "f81366e5-fbdb-4a0b-a91f-f603d9b8570d" (UID: "f81366e5-fbdb-4a0b-a91f-f603d9b8570d"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.801495 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "f81366e5-fbdb-4a0b-a91f-f603d9b8570d" (UID: "f81366e5-fbdb-4a0b-a91f-f603d9b8570d"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.801682 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-kube-api-access-7b5rg" (OuterVolumeSpecName: "kube-api-access-7b5rg") pod "f81366e5-fbdb-4a0b-a91f-f603d9b8570d" (UID: "f81366e5-fbdb-4a0b-a91f-f603d9b8570d"). InnerVolumeSpecName "kube-api-access-7b5rg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.802277 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "f81366e5-fbdb-4a0b-a91f-f603d9b8570d" (UID: "f81366e5-fbdb-4a0b-a91f-f603d9b8570d"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.891558 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct\" (UID: \"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.891656 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcj8g\" (UniqueName: \"kubernetes.io/projected/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-kube-api-access-lcj8g\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct\" (UID: \"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.892006 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct\" (UID: \"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.892150 4948 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.892247 4948 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.892274 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7b5rg\" (UniqueName: \"kubernetes.io/projected/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-kube-api-access-7b5rg\") on node \"crc\" DevicePath \"\"" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.892342 4948 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.892375 4948 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.892465 4948 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-service-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.892546 4948 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.892394 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct\" (UID: \"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.892410 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct\" (UID: \"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.914180 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcj8g\" (UniqueName: \"kubernetes.io/projected/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-kube-api-access-lcj8g\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct\" (UID: \"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.945367 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" Feb 20 08:17:06 crc kubenswrapper[4948]: I0220 08:17:06.147253 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct"] Feb 20 08:17:06 crc kubenswrapper[4948]: W0220 08:17:06.156401 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e5e2f81_8ea9_4ddb_b2ab_584c40b15916.slice/crio-b58d77869abc6a06a40c2616d2fffd5dda5671eb4c1eccfaea6ec70a3e40332b WatchSource:0}: Error finding container b58d77869abc6a06a40c2616d2fffd5dda5671eb4c1eccfaea6ec70a3e40332b: Status 404 returned error can't find the container with id b58d77869abc6a06a40c2616d2fffd5dda5671eb4c1eccfaea6ec70a3e40332b Feb 20 08:17:06 crc kubenswrapper[4948]: I0220 08:17:06.349266 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-k8zcr_f81366e5-fbdb-4a0b-a91f-f603d9b8570d/console/0.log" Feb 20 08:17:06 crc kubenswrapper[4948]: I0220 08:17:06.349351 4948 generic.go:334] "Generic (PLEG): container finished" podID="f81366e5-fbdb-4a0b-a91f-f603d9b8570d" containerID="3745ecb9e997c9c1a842fb71042af1df4c9828447494e128f09e0903f609bd78" exitCode=2 Feb 20 08:17:06 crc kubenswrapper[4948]: I0220 08:17:06.349449 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-k8zcr" event={"ID":"f81366e5-fbdb-4a0b-a91f-f603d9b8570d","Type":"ContainerDied","Data":"3745ecb9e997c9c1a842fb71042af1df4c9828447494e128f09e0903f609bd78"} Feb 20 08:17:06 crc kubenswrapper[4948]: I0220 08:17:06.349493 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-k8zcr" event={"ID":"f81366e5-fbdb-4a0b-a91f-f603d9b8570d","Type":"ContainerDied","Data":"61712307e24396811f0211771ba686b9bbc2eb0844719329e63b16d44d321de6"} Feb 20 08:17:06 crc kubenswrapper[4948]: I0220 08:17:06.349530 4948 scope.go:117] "RemoveContainer" containerID="3745ecb9e997c9c1a842fb71042af1df4c9828447494e128f09e0903f609bd78" Feb 20 08:17:06 crc kubenswrapper[4948]: I0220 08:17:06.349570 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:17:06 crc kubenswrapper[4948]: I0220 08:17:06.353281 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" event={"ID":"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916","Type":"ContainerStarted","Data":"5be15e1279406c468ca430c5a05ee089f20fb99bdd79d38237b6ed24ccf2f18c"} Feb 20 08:17:06 crc kubenswrapper[4948]: I0220 08:17:06.353350 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" event={"ID":"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916","Type":"ContainerStarted","Data":"b58d77869abc6a06a40c2616d2fffd5dda5671eb4c1eccfaea6ec70a3e40332b"} Feb 20 08:17:06 crc kubenswrapper[4948]: I0220 08:17:06.390548 4948 scope.go:117] "RemoveContainer" containerID="3745ecb9e997c9c1a842fb71042af1df4c9828447494e128f09e0903f609bd78" Feb 20 08:17:06 crc kubenswrapper[4948]: E0220 08:17:06.391517 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3745ecb9e997c9c1a842fb71042af1df4c9828447494e128f09e0903f609bd78\": container with ID starting with 3745ecb9e997c9c1a842fb71042af1df4c9828447494e128f09e0903f609bd78 not found: ID does not exist" containerID="3745ecb9e997c9c1a842fb71042af1df4c9828447494e128f09e0903f609bd78" Feb 20 08:17:06 crc kubenswrapper[4948]: I0220 08:17:06.391595 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3745ecb9e997c9c1a842fb71042af1df4c9828447494e128f09e0903f609bd78"} err="failed to get container status \"3745ecb9e997c9c1a842fb71042af1df4c9828447494e128f09e0903f609bd78\": rpc error: code = NotFound desc = could not find container \"3745ecb9e997c9c1a842fb71042af1df4c9828447494e128f09e0903f609bd78\": container with ID starting with 3745ecb9e997c9c1a842fb71042af1df4c9828447494e128f09e0903f609bd78 not found: ID does not exist" Feb 20 08:17:06 crc kubenswrapper[4948]: I0220 08:17:06.433209 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-k8zcr"] Feb 20 08:17:06 crc kubenswrapper[4948]: I0220 08:17:06.435588 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-k8zcr"] Feb 20 08:17:07 crc kubenswrapper[4948]: I0220 08:17:07.367371 4948 generic.go:334] "Generic (PLEG): container finished" podID="9e5e2f81-8ea9-4ddb-b2ab-584c40b15916" containerID="5be15e1279406c468ca430c5a05ee089f20fb99bdd79d38237b6ed24ccf2f18c" exitCode=0 Feb 20 08:17:07 crc kubenswrapper[4948]: I0220 08:17:07.367471 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" event={"ID":"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916","Type":"ContainerDied","Data":"5be15e1279406c468ca430c5a05ee089f20fb99bdd79d38237b6ed24ccf2f18c"} Feb 20 08:17:07 crc kubenswrapper[4948]: I0220 08:17:07.736770 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f81366e5-fbdb-4a0b-a91f-f603d9b8570d" path="/var/lib/kubelet/pods/f81366e5-fbdb-4a0b-a91f-f603d9b8570d/volumes" Feb 20 08:17:09 crc kubenswrapper[4948]: I0220 08:17:09.386143 4948 generic.go:334] "Generic (PLEG): container finished" podID="9e5e2f81-8ea9-4ddb-b2ab-584c40b15916" containerID="dfd6183de2e06d2690aa7e1da7e9625df33f176428e6c2492950a7e9940ee6df" exitCode=0 Feb 20 08:17:09 crc kubenswrapper[4948]: I0220 08:17:09.386216 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" event={"ID":"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916","Type":"ContainerDied","Data":"dfd6183de2e06d2690aa7e1da7e9625df33f176428e6c2492950a7e9940ee6df"} Feb 20 08:17:10 crc kubenswrapper[4948]: I0220 08:17:10.397312 4948 generic.go:334] "Generic (PLEG): container finished" podID="9e5e2f81-8ea9-4ddb-b2ab-584c40b15916" containerID="cfec89ba0a0c86027899432b1991c477117822e794036be26b25129aec08acc2" exitCode=0 Feb 20 08:17:10 crc kubenswrapper[4948]: I0220 08:17:10.397378 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" event={"ID":"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916","Type":"ContainerDied","Data":"cfec89ba0a0c86027899432b1991c477117822e794036be26b25129aec08acc2"} Feb 20 08:17:11 crc kubenswrapper[4948]: I0220 08:17:11.753541 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" Feb 20 08:17:11 crc kubenswrapper[4948]: I0220 08:17:11.879883 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcj8g\" (UniqueName: \"kubernetes.io/projected/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-kube-api-access-lcj8g\") pod \"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916\" (UID: \"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916\") " Feb 20 08:17:11 crc kubenswrapper[4948]: I0220 08:17:11.880065 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-bundle\") pod \"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916\" (UID: \"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916\") " Feb 20 08:17:11 crc kubenswrapper[4948]: I0220 08:17:11.880193 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-util\") pod \"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916\" (UID: \"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916\") " Feb 20 08:17:11 crc kubenswrapper[4948]: I0220 08:17:11.882197 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-bundle" (OuterVolumeSpecName: "bundle") pod "9e5e2f81-8ea9-4ddb-b2ab-584c40b15916" (UID: "9e5e2f81-8ea9-4ddb-b2ab-584c40b15916"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:17:11 crc kubenswrapper[4948]: I0220 08:17:11.890144 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-kube-api-access-lcj8g" (OuterVolumeSpecName: "kube-api-access-lcj8g") pod "9e5e2f81-8ea9-4ddb-b2ab-584c40b15916" (UID: "9e5e2f81-8ea9-4ddb-b2ab-584c40b15916"). InnerVolumeSpecName "kube-api-access-lcj8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:17:11 crc kubenswrapper[4948]: I0220 08:17:11.905220 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-util" (OuterVolumeSpecName: "util") pod "9e5e2f81-8ea9-4ddb-b2ab-584c40b15916" (UID: "9e5e2f81-8ea9-4ddb-b2ab-584c40b15916"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:17:11 crc kubenswrapper[4948]: I0220 08:17:11.981815 4948 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:17:11 crc kubenswrapper[4948]: I0220 08:17:11.981907 4948 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-util\") on node \"crc\" DevicePath \"\"" Feb 20 08:17:11 crc kubenswrapper[4948]: I0220 08:17:11.981962 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcj8g\" (UniqueName: \"kubernetes.io/projected/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-kube-api-access-lcj8g\") on node \"crc\" DevicePath \"\"" Feb 20 08:17:12 crc kubenswrapper[4948]: I0220 08:17:12.415143 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" event={"ID":"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916","Type":"ContainerDied","Data":"b58d77869abc6a06a40c2616d2fffd5dda5671eb4c1eccfaea6ec70a3e40332b"} Feb 20 08:17:12 crc kubenswrapper[4948]: I0220 08:17:12.415203 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b58d77869abc6a06a40c2616d2fffd5dda5671eb4c1eccfaea6ec70a3e40332b" Feb 20 08:17:12 crc kubenswrapper[4948]: I0220 08:17:12.415261 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.582734 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs"] Feb 20 08:17:20 crc kubenswrapper[4948]: E0220 08:17:20.583515 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e5e2f81-8ea9-4ddb-b2ab-584c40b15916" containerName="extract" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.583527 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e5e2f81-8ea9-4ddb-b2ab-584c40b15916" containerName="extract" Feb 20 08:17:20 crc kubenswrapper[4948]: E0220 08:17:20.583543 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e5e2f81-8ea9-4ddb-b2ab-584c40b15916" containerName="pull" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.583549 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e5e2f81-8ea9-4ddb-b2ab-584c40b15916" containerName="pull" Feb 20 08:17:20 crc kubenswrapper[4948]: E0220 08:17:20.583559 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e5e2f81-8ea9-4ddb-b2ab-584c40b15916" containerName="util" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.583565 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e5e2f81-8ea9-4ddb-b2ab-584c40b15916" containerName="util" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.583665 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e5e2f81-8ea9-4ddb-b2ab-584c40b15916" containerName="extract" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.584055 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.587549 4948 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.587654 4948 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.587713 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.587851 4948 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-pqnxf" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.587860 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.603193 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs"] Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.696127 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dncwx\" (UniqueName: \"kubernetes.io/projected/b4f86a2e-15f9-441c-953b-49f331d4122e-kube-api-access-dncwx\") pod \"metallb-operator-controller-manager-8478c6b8cc-v98zs\" (UID: \"b4f86a2e-15f9-441c-953b-49f331d4122e\") " pod="metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.696182 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b4f86a2e-15f9-441c-953b-49f331d4122e-apiservice-cert\") pod \"metallb-operator-controller-manager-8478c6b8cc-v98zs\" (UID: \"b4f86a2e-15f9-441c-953b-49f331d4122e\") " pod="metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.696273 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b4f86a2e-15f9-441c-953b-49f331d4122e-webhook-cert\") pod \"metallb-operator-controller-manager-8478c6b8cc-v98zs\" (UID: \"b4f86a2e-15f9-441c-953b-49f331d4122e\") " pod="metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.797410 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dncwx\" (UniqueName: \"kubernetes.io/projected/b4f86a2e-15f9-441c-953b-49f331d4122e-kube-api-access-dncwx\") pod \"metallb-operator-controller-manager-8478c6b8cc-v98zs\" (UID: \"b4f86a2e-15f9-441c-953b-49f331d4122e\") " pod="metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.797468 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b4f86a2e-15f9-441c-953b-49f331d4122e-apiservice-cert\") pod \"metallb-operator-controller-manager-8478c6b8cc-v98zs\" (UID: \"b4f86a2e-15f9-441c-953b-49f331d4122e\") " pod="metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.797741 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b4f86a2e-15f9-441c-953b-49f331d4122e-webhook-cert\") pod \"metallb-operator-controller-manager-8478c6b8cc-v98zs\" (UID: \"b4f86a2e-15f9-441c-953b-49f331d4122e\") " pod="metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.803749 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b4f86a2e-15f9-441c-953b-49f331d4122e-apiservice-cert\") pod \"metallb-operator-controller-manager-8478c6b8cc-v98zs\" (UID: \"b4f86a2e-15f9-441c-953b-49f331d4122e\") " pod="metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.803808 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b4f86a2e-15f9-441c-953b-49f331d4122e-webhook-cert\") pod \"metallb-operator-controller-manager-8478c6b8cc-v98zs\" (UID: \"b4f86a2e-15f9-441c-953b-49f331d4122e\") " pod="metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.825525 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dncwx\" (UniqueName: \"kubernetes.io/projected/b4f86a2e-15f9-441c-953b-49f331d4122e-kube-api-access-dncwx\") pod \"metallb-operator-controller-manager-8478c6b8cc-v98zs\" (UID: \"b4f86a2e-15f9-441c-953b-49f331d4122e\") " pod="metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.899759 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.915074 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs"] Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.916114 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.919633 4948 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.923139 4948 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.930244 4948 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-qc469" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.951916 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs"] Feb 20 08:17:21 crc kubenswrapper[4948]: I0220 08:17:21.000398 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dz9g7\" (UniqueName: \"kubernetes.io/projected/07022b13-7dc1-45d1-a626-91792bd3aa90-kube-api-access-dz9g7\") pod \"metallb-operator-webhook-server-5fcd7d546c-kgpxs\" (UID: \"07022b13-7dc1-45d1-a626-91792bd3aa90\") " pod="metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs" Feb 20 08:17:21 crc kubenswrapper[4948]: I0220 08:17:21.000472 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/07022b13-7dc1-45d1-a626-91792bd3aa90-apiservice-cert\") pod \"metallb-operator-webhook-server-5fcd7d546c-kgpxs\" (UID: \"07022b13-7dc1-45d1-a626-91792bd3aa90\") " pod="metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs" Feb 20 08:17:21 crc kubenswrapper[4948]: I0220 08:17:21.000602 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/07022b13-7dc1-45d1-a626-91792bd3aa90-webhook-cert\") pod \"metallb-operator-webhook-server-5fcd7d546c-kgpxs\" (UID: \"07022b13-7dc1-45d1-a626-91792bd3aa90\") " pod="metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs" Feb 20 08:17:21 crc kubenswrapper[4948]: I0220 08:17:21.101851 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/07022b13-7dc1-45d1-a626-91792bd3aa90-webhook-cert\") pod \"metallb-operator-webhook-server-5fcd7d546c-kgpxs\" (UID: \"07022b13-7dc1-45d1-a626-91792bd3aa90\") " pod="metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs" Feb 20 08:17:21 crc kubenswrapper[4948]: I0220 08:17:21.102309 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dz9g7\" (UniqueName: \"kubernetes.io/projected/07022b13-7dc1-45d1-a626-91792bd3aa90-kube-api-access-dz9g7\") pod \"metallb-operator-webhook-server-5fcd7d546c-kgpxs\" (UID: \"07022b13-7dc1-45d1-a626-91792bd3aa90\") " pod="metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs" Feb 20 08:17:21 crc kubenswrapper[4948]: I0220 08:17:21.102354 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/07022b13-7dc1-45d1-a626-91792bd3aa90-apiservice-cert\") pod \"metallb-operator-webhook-server-5fcd7d546c-kgpxs\" (UID: \"07022b13-7dc1-45d1-a626-91792bd3aa90\") " pod="metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs" Feb 20 08:17:21 crc kubenswrapper[4948]: I0220 08:17:21.107074 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/07022b13-7dc1-45d1-a626-91792bd3aa90-apiservice-cert\") pod \"metallb-operator-webhook-server-5fcd7d546c-kgpxs\" (UID: \"07022b13-7dc1-45d1-a626-91792bd3aa90\") " pod="metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs" Feb 20 08:17:21 crc kubenswrapper[4948]: I0220 08:17:21.121181 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/07022b13-7dc1-45d1-a626-91792bd3aa90-webhook-cert\") pod \"metallb-operator-webhook-server-5fcd7d546c-kgpxs\" (UID: \"07022b13-7dc1-45d1-a626-91792bd3aa90\") " pod="metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs" Feb 20 08:17:21 crc kubenswrapper[4948]: I0220 08:17:21.121278 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dz9g7\" (UniqueName: \"kubernetes.io/projected/07022b13-7dc1-45d1-a626-91792bd3aa90-kube-api-access-dz9g7\") pod \"metallb-operator-webhook-server-5fcd7d546c-kgpxs\" (UID: \"07022b13-7dc1-45d1-a626-91792bd3aa90\") " pod="metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs" Feb 20 08:17:21 crc kubenswrapper[4948]: I0220 08:17:21.147585 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs"] Feb 20 08:17:21 crc kubenswrapper[4948]: W0220 08:17:21.154745 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4f86a2e_15f9_441c_953b_49f331d4122e.slice/crio-fa52bf8c7effb4b51c609222f52518944307009826752b172359108a4af822c0 WatchSource:0}: Error finding container fa52bf8c7effb4b51c609222f52518944307009826752b172359108a4af822c0: Status 404 returned error can't find the container with id fa52bf8c7effb4b51c609222f52518944307009826752b172359108a4af822c0 Feb 20 08:17:21 crc kubenswrapper[4948]: I0220 08:17:21.285261 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs" Feb 20 08:17:21 crc kubenswrapper[4948]: I0220 08:17:21.468779 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs" event={"ID":"b4f86a2e-15f9-441c-953b-49f331d4122e","Type":"ContainerStarted","Data":"fa52bf8c7effb4b51c609222f52518944307009826752b172359108a4af822c0"} Feb 20 08:17:21 crc kubenswrapper[4948]: I0220 08:17:21.499220 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs"] Feb 20 08:17:21 crc kubenswrapper[4948]: W0220 08:17:21.506015 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod07022b13_7dc1_45d1_a626_91792bd3aa90.slice/crio-6500653b5f55a54ab5eb26076b35eb76477dc2e1b4770013d2121d1bf1f20b85 WatchSource:0}: Error finding container 6500653b5f55a54ab5eb26076b35eb76477dc2e1b4770013d2121d1bf1f20b85: Status 404 returned error can't find the container with id 6500653b5f55a54ab5eb26076b35eb76477dc2e1b4770013d2121d1bf1f20b85 Feb 20 08:17:22 crc kubenswrapper[4948]: I0220 08:17:22.474680 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs" event={"ID":"07022b13-7dc1-45d1-a626-91792bd3aa90","Type":"ContainerStarted","Data":"6500653b5f55a54ab5eb26076b35eb76477dc2e1b4770013d2121d1bf1f20b85"} Feb 20 08:17:26 crc kubenswrapper[4948]: I0220 08:17:26.497914 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs" event={"ID":"07022b13-7dc1-45d1-a626-91792bd3aa90","Type":"ContainerStarted","Data":"841867153dd83886e9618e9a4fae6c87b91836589203878bc1a8db173192ca25"} Feb 20 08:17:26 crc kubenswrapper[4948]: I0220 08:17:26.500303 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs" event={"ID":"b4f86a2e-15f9-441c-953b-49f331d4122e","Type":"ContainerStarted","Data":"ab6d5674d550742387ada1d66fe00576345be4280789d0114664533ce4ca4a49"} Feb 20 08:17:26 crc kubenswrapper[4948]: I0220 08:17:26.500459 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs" Feb 20 08:17:26 crc kubenswrapper[4948]: I0220 08:17:26.522012 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs" podStartSLOduration=2.127778239 podStartE2EDuration="6.521985757s" podCreationTimestamp="2026-02-20 08:17:20 +0000 UTC" firstStartedPulling="2026-02-20 08:17:21.509822408 +0000 UTC m=+690.484317228" lastFinishedPulling="2026-02-20 08:17:25.904029926 +0000 UTC m=+694.878524746" observedRunningTime="2026-02-20 08:17:26.516051128 +0000 UTC m=+695.490545988" watchObservedRunningTime="2026-02-20 08:17:26.521985757 +0000 UTC m=+695.496480587" Feb 20 08:17:26 crc kubenswrapper[4948]: I0220 08:17:26.548203 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs" podStartSLOduration=1.891160094 podStartE2EDuration="6.548183902s" podCreationTimestamp="2026-02-20 08:17:20 +0000 UTC" firstStartedPulling="2026-02-20 08:17:21.15667996 +0000 UTC m=+690.131174780" lastFinishedPulling="2026-02-20 08:17:25.813703768 +0000 UTC m=+694.788198588" observedRunningTime="2026-02-20 08:17:26.54532811 +0000 UTC m=+695.519822930" watchObservedRunningTime="2026-02-20 08:17:26.548183902 +0000 UTC m=+695.522678732" Feb 20 08:17:27 crc kubenswrapper[4948]: I0220 08:17:27.507158 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs" Feb 20 08:17:38 crc kubenswrapper[4948]: I0220 08:17:38.024801 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:17:38 crc kubenswrapper[4948]: I0220 08:17:38.027058 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:17:41 crc kubenswrapper[4948]: I0220 08:17:41.296542 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs" Feb 20 08:18:00 crc kubenswrapper[4948]: I0220 08:18:00.903364 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.592153 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-k66nh"] Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.595525 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.600493 4948 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.600507 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.601319 4948 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-xnrzs" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.616676 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-7zsxs"] Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.616964 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0131d05b-619d-43d8-aa38-ad6e9ce52e11-metrics-certs\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.617058 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/0131d05b-619d-43d8-aa38-ad6e9ce52e11-reloader\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.617099 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/0131d05b-619d-43d8-aa38-ad6e9ce52e11-frr-startup\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.617136 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/0131d05b-619d-43d8-aa38-ad6e9ce52e11-frr-conf\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.617240 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/0131d05b-619d-43d8-aa38-ad6e9ce52e11-metrics\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.617297 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/0131d05b-619d-43d8-aa38-ad6e9ce52e11-frr-sockets\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.617610 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q765x\" (UniqueName: \"kubernetes.io/projected/0131d05b-619d-43d8-aa38-ad6e9ce52e11-kube-api-access-q765x\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.617866 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7zsxs" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.619878 4948 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.635215 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-7zsxs"] Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.680748 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-h7xhj"] Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.681641 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-h7xhj" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.683320 4948 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.684462 4948 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.684722 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.684900 4948 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-fpnk6" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.713418 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-69bbfbf88f-c6sqk"] Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.714177 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-69bbfbf88f-c6sqk" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.716412 4948 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.718455 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/2c199acd-2fd3-4b9f-b50c-a266191c4777-memberlist\") pod \"speaker-h7xhj\" (UID: \"2c199acd-2fd3-4b9f-b50c-a266191c4777\") " pod="metallb-system/speaker-h7xhj" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.718497 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/2c199acd-2fd3-4b9f-b50c-a266191c4777-metallb-excludel2\") pod \"speaker-h7xhj\" (UID: \"2c199acd-2fd3-4b9f-b50c-a266191c4777\") " pod="metallb-system/speaker-h7xhj" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.718527 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/0131d05b-619d-43d8-aa38-ad6e9ce52e11-metrics\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.718555 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/0131d05b-619d-43d8-aa38-ad6e9ce52e11-frr-sockets\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.718607 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2506d4e2-b434-4fe4-970e-7cd14601677d-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-7zsxs\" (UID: \"2506d4e2-b434-4fe4-970e-7cd14601677d\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7zsxs" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.719138 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/0131d05b-619d-43d8-aa38-ad6e9ce52e11-metrics\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.719159 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q765x\" (UniqueName: \"kubernetes.io/projected/0131d05b-619d-43d8-aa38-ad6e9ce52e11-kube-api-access-q765x\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.719160 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/0131d05b-619d-43d8-aa38-ad6e9ce52e11-frr-sockets\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.719239 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0131d05b-619d-43d8-aa38-ad6e9ce52e11-metrics-certs\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.719291 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/0131d05b-619d-43d8-aa38-ad6e9ce52e11-reloader\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.719327 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2c199acd-2fd3-4b9f-b50c-a266191c4777-metrics-certs\") pod \"speaker-h7xhj\" (UID: \"2c199acd-2fd3-4b9f-b50c-a266191c4777\") " pod="metallb-system/speaker-h7xhj" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.719361 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/0131d05b-619d-43d8-aa38-ad6e9ce52e11-frr-startup\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.719391 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/0131d05b-619d-43d8-aa38-ad6e9ce52e11-frr-conf\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: E0220 08:18:01.719591 4948 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.719659 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/0131d05b-619d-43d8-aa38-ad6e9ce52e11-reloader\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.719781 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/0131d05b-619d-43d8-aa38-ad6e9ce52e11-frr-conf\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: E0220 08:18:01.719892 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0131d05b-619d-43d8-aa38-ad6e9ce52e11-metrics-certs podName:0131d05b-619d-43d8-aa38-ad6e9ce52e11 nodeName:}" failed. No retries permitted until 2026-02-20 08:18:02.219872578 +0000 UTC m=+731.194367398 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0131d05b-619d-43d8-aa38-ad6e9ce52e11-metrics-certs") pod "frr-k8s-k66nh" (UID: "0131d05b-619d-43d8-aa38-ad6e9ce52e11") : secret "frr-k8s-certs-secret" not found Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.719922 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqst8\" (UniqueName: \"kubernetes.io/projected/2c199acd-2fd3-4b9f-b50c-a266191c4777-kube-api-access-dqst8\") pod \"speaker-h7xhj\" (UID: \"2c199acd-2fd3-4b9f-b50c-a266191c4777\") " pod="metallb-system/speaker-h7xhj" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.720005 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgh5n\" (UniqueName: \"kubernetes.io/projected/2506d4e2-b434-4fe4-970e-7cd14601677d-kube-api-access-wgh5n\") pod \"frr-k8s-webhook-server-78b44bf5bb-7zsxs\" (UID: \"2506d4e2-b434-4fe4-970e-7cd14601677d\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7zsxs" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.720901 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/0131d05b-619d-43d8-aa38-ad6e9ce52e11-frr-startup\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.732457 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-69bbfbf88f-c6sqk"] Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.761208 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q765x\" (UniqueName: \"kubernetes.io/projected/0131d05b-619d-43d8-aa38-ad6e9ce52e11-kube-api-access-q765x\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.820890 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2c199acd-2fd3-4b9f-b50c-a266191c4777-metrics-certs\") pod \"speaker-h7xhj\" (UID: \"2c199acd-2fd3-4b9f-b50c-a266191c4777\") " pod="metallb-system/speaker-h7xhj" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.820948 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqst8\" (UniqueName: \"kubernetes.io/projected/2c199acd-2fd3-4b9f-b50c-a266191c4777-kube-api-access-dqst8\") pod \"speaker-h7xhj\" (UID: \"2c199acd-2fd3-4b9f-b50c-a266191c4777\") " pod="metallb-system/speaker-h7xhj" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.820975 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/68ee0185-13d4-45c7-9a39-bb474e9314d0-cert\") pod \"controller-69bbfbf88f-c6sqk\" (UID: \"68ee0185-13d4-45c7-9a39-bb474e9314d0\") " pod="metallb-system/controller-69bbfbf88f-c6sqk" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.821011 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgh5n\" (UniqueName: \"kubernetes.io/projected/2506d4e2-b434-4fe4-970e-7cd14601677d-kube-api-access-wgh5n\") pod \"frr-k8s-webhook-server-78b44bf5bb-7zsxs\" (UID: \"2506d4e2-b434-4fe4-970e-7cd14601677d\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7zsxs" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.821029 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/2c199acd-2fd3-4b9f-b50c-a266191c4777-memberlist\") pod \"speaker-h7xhj\" (UID: \"2c199acd-2fd3-4b9f-b50c-a266191c4777\") " pod="metallb-system/speaker-h7xhj" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.821046 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/2c199acd-2fd3-4b9f-b50c-a266191c4777-metallb-excludel2\") pod \"speaker-h7xhj\" (UID: \"2c199acd-2fd3-4b9f-b50c-a266191c4777\") " pod="metallb-system/speaker-h7xhj" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.821066 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/68ee0185-13d4-45c7-9a39-bb474e9314d0-metrics-certs\") pod \"controller-69bbfbf88f-c6sqk\" (UID: \"68ee0185-13d4-45c7-9a39-bb474e9314d0\") " pod="metallb-system/controller-69bbfbf88f-c6sqk" Feb 20 08:18:01 crc kubenswrapper[4948]: E0220 08:18:01.821066 4948 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.821100 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2506d4e2-b434-4fe4-970e-7cd14601677d-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-7zsxs\" (UID: \"2506d4e2-b434-4fe4-970e-7cd14601677d\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7zsxs" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.821123 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxtfc\" (UniqueName: \"kubernetes.io/projected/68ee0185-13d4-45c7-9a39-bb474e9314d0-kube-api-access-qxtfc\") pod \"controller-69bbfbf88f-c6sqk\" (UID: \"68ee0185-13d4-45c7-9a39-bb474e9314d0\") " pod="metallb-system/controller-69bbfbf88f-c6sqk" Feb 20 08:18:01 crc kubenswrapper[4948]: E0220 08:18:01.821137 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2c199acd-2fd3-4b9f-b50c-a266191c4777-metrics-certs podName:2c199acd-2fd3-4b9f-b50c-a266191c4777 nodeName:}" failed. No retries permitted until 2026-02-20 08:18:02.32111968 +0000 UTC m=+731.295614500 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2c199acd-2fd3-4b9f-b50c-a266191c4777-metrics-certs") pod "speaker-h7xhj" (UID: "2c199acd-2fd3-4b9f-b50c-a266191c4777") : secret "speaker-certs-secret" not found Feb 20 08:18:01 crc kubenswrapper[4948]: E0220 08:18:01.821520 4948 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 20 08:18:01 crc kubenswrapper[4948]: E0220 08:18:01.821569 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2c199acd-2fd3-4b9f-b50c-a266191c4777-memberlist podName:2c199acd-2fd3-4b9f-b50c-a266191c4777 nodeName:}" failed. No retries permitted until 2026-02-20 08:18:02.32155485 +0000 UTC m=+731.296049670 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/2c199acd-2fd3-4b9f-b50c-a266191c4777-memberlist") pod "speaker-h7xhj" (UID: "2c199acd-2fd3-4b9f-b50c-a266191c4777") : secret "metallb-memberlist" not found Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.821702 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/2c199acd-2fd3-4b9f-b50c-a266191c4777-metallb-excludel2\") pod \"speaker-h7xhj\" (UID: \"2c199acd-2fd3-4b9f-b50c-a266191c4777\") " pod="metallb-system/speaker-h7xhj" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.826212 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2506d4e2-b434-4fe4-970e-7cd14601677d-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-7zsxs\" (UID: \"2506d4e2-b434-4fe4-970e-7cd14601677d\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7zsxs" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.842506 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqst8\" (UniqueName: \"kubernetes.io/projected/2c199acd-2fd3-4b9f-b50c-a266191c4777-kube-api-access-dqst8\") pod \"speaker-h7xhj\" (UID: \"2c199acd-2fd3-4b9f-b50c-a266191c4777\") " pod="metallb-system/speaker-h7xhj" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.853742 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgh5n\" (UniqueName: \"kubernetes.io/projected/2506d4e2-b434-4fe4-970e-7cd14601677d-kube-api-access-wgh5n\") pod \"frr-k8s-webhook-server-78b44bf5bb-7zsxs\" (UID: \"2506d4e2-b434-4fe4-970e-7cd14601677d\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7zsxs" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.922031 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/68ee0185-13d4-45c7-9a39-bb474e9314d0-cert\") pod \"controller-69bbfbf88f-c6sqk\" (UID: \"68ee0185-13d4-45c7-9a39-bb474e9314d0\") " pod="metallb-system/controller-69bbfbf88f-c6sqk" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.922722 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/68ee0185-13d4-45c7-9a39-bb474e9314d0-metrics-certs\") pod \"controller-69bbfbf88f-c6sqk\" (UID: \"68ee0185-13d4-45c7-9a39-bb474e9314d0\") " pod="metallb-system/controller-69bbfbf88f-c6sqk" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.922900 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxtfc\" (UniqueName: \"kubernetes.io/projected/68ee0185-13d4-45c7-9a39-bb474e9314d0-kube-api-access-qxtfc\") pod \"controller-69bbfbf88f-c6sqk\" (UID: \"68ee0185-13d4-45c7-9a39-bb474e9314d0\") " pod="metallb-system/controller-69bbfbf88f-c6sqk" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.925056 4948 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.926650 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/68ee0185-13d4-45c7-9a39-bb474e9314d0-metrics-certs\") pod \"controller-69bbfbf88f-c6sqk\" (UID: \"68ee0185-13d4-45c7-9a39-bb474e9314d0\") " pod="metallb-system/controller-69bbfbf88f-c6sqk" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.935525 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/68ee0185-13d4-45c7-9a39-bb474e9314d0-cert\") pod \"controller-69bbfbf88f-c6sqk\" (UID: \"68ee0185-13d4-45c7-9a39-bb474e9314d0\") " pod="metallb-system/controller-69bbfbf88f-c6sqk" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.938506 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7zsxs" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.955015 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxtfc\" (UniqueName: \"kubernetes.io/projected/68ee0185-13d4-45c7-9a39-bb474e9314d0-kube-api-access-qxtfc\") pod \"controller-69bbfbf88f-c6sqk\" (UID: \"68ee0185-13d4-45c7-9a39-bb474e9314d0\") " pod="metallb-system/controller-69bbfbf88f-c6sqk" Feb 20 08:18:02 crc kubenswrapper[4948]: I0220 08:18:02.026507 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-69bbfbf88f-c6sqk" Feb 20 08:18:02 crc kubenswrapper[4948]: I0220 08:18:02.227070 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0131d05b-619d-43d8-aa38-ad6e9ce52e11-metrics-certs\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:02 crc kubenswrapper[4948]: I0220 08:18:02.235070 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0131d05b-619d-43d8-aa38-ad6e9ce52e11-metrics-certs\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:02 crc kubenswrapper[4948]: I0220 08:18:02.329550 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2c199acd-2fd3-4b9f-b50c-a266191c4777-metrics-certs\") pod \"speaker-h7xhj\" (UID: \"2c199acd-2fd3-4b9f-b50c-a266191c4777\") " pod="metallb-system/speaker-h7xhj" Feb 20 08:18:02 crc kubenswrapper[4948]: I0220 08:18:02.329736 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/2c199acd-2fd3-4b9f-b50c-a266191c4777-memberlist\") pod \"speaker-h7xhj\" (UID: \"2c199acd-2fd3-4b9f-b50c-a266191c4777\") " pod="metallb-system/speaker-h7xhj" Feb 20 08:18:02 crc kubenswrapper[4948]: E0220 08:18:02.329903 4948 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 20 08:18:02 crc kubenswrapper[4948]: E0220 08:18:02.330006 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2c199acd-2fd3-4b9f-b50c-a266191c4777-memberlist podName:2c199acd-2fd3-4b9f-b50c-a266191c4777 nodeName:}" failed. No retries permitted until 2026-02-20 08:18:03.329952892 +0000 UTC m=+732.304447742 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/2c199acd-2fd3-4b9f-b50c-a266191c4777-memberlist") pod "speaker-h7xhj" (UID: "2c199acd-2fd3-4b9f-b50c-a266191c4777") : secret "metallb-memberlist" not found Feb 20 08:18:02 crc kubenswrapper[4948]: I0220 08:18:02.332704 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2c199acd-2fd3-4b9f-b50c-a266191c4777-metrics-certs\") pod \"speaker-h7xhj\" (UID: \"2c199acd-2fd3-4b9f-b50c-a266191c4777\") " pod="metallb-system/speaker-h7xhj" Feb 20 08:18:02 crc kubenswrapper[4948]: I0220 08:18:02.401675 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-7zsxs"] Feb 20 08:18:02 crc kubenswrapper[4948]: I0220 08:18:02.470596 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-69bbfbf88f-c6sqk"] Feb 20 08:18:02 crc kubenswrapper[4948]: W0220 08:18:02.483344 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod68ee0185_13d4_45c7_9a39_bb474e9314d0.slice/crio-29ad037d86595ebeca4c463306356d180d574a77f75e87dfb0da6a250144e832 WatchSource:0}: Error finding container 29ad037d86595ebeca4c463306356d180d574a77f75e87dfb0da6a250144e832: Status 404 returned error can't find the container with id 29ad037d86595ebeca4c463306356d180d574a77f75e87dfb0da6a250144e832 Feb 20 08:18:02 crc kubenswrapper[4948]: I0220 08:18:02.518020 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:02 crc kubenswrapper[4948]: I0220 08:18:02.742240 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7zsxs" event={"ID":"2506d4e2-b434-4fe4-970e-7cd14601677d","Type":"ContainerStarted","Data":"374f01b7e30c3730f2d0fe4500f1a90d86e1a9243240a4cc2a39a2e1093752ac"} Feb 20 08:18:02 crc kubenswrapper[4948]: I0220 08:18:02.743173 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-k66nh" event={"ID":"0131d05b-619d-43d8-aa38-ad6e9ce52e11","Type":"ContainerStarted","Data":"c12d45fc8456567f3dfc5e086c5767ad4130f5148db37efe222ca6aa614aa44f"} Feb 20 08:18:02 crc kubenswrapper[4948]: I0220 08:18:02.753195 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-c6sqk" event={"ID":"68ee0185-13d4-45c7-9a39-bb474e9314d0","Type":"ContainerStarted","Data":"ec977772b2186974da9e5022ec788dee392cac9a842ebd0f652092982d64289e"} Feb 20 08:18:02 crc kubenswrapper[4948]: I0220 08:18:02.753235 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-c6sqk" event={"ID":"68ee0185-13d4-45c7-9a39-bb474e9314d0","Type":"ContainerStarted","Data":"29ad037d86595ebeca4c463306356d180d574a77f75e87dfb0da6a250144e832"} Feb 20 08:18:03 crc kubenswrapper[4948]: I0220 08:18:03.348541 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/2c199acd-2fd3-4b9f-b50c-a266191c4777-memberlist\") pod \"speaker-h7xhj\" (UID: \"2c199acd-2fd3-4b9f-b50c-a266191c4777\") " pod="metallb-system/speaker-h7xhj" Feb 20 08:18:03 crc kubenswrapper[4948]: I0220 08:18:03.365008 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/2c199acd-2fd3-4b9f-b50c-a266191c4777-memberlist\") pod \"speaker-h7xhj\" (UID: \"2c199acd-2fd3-4b9f-b50c-a266191c4777\") " pod="metallb-system/speaker-h7xhj" Feb 20 08:18:03 crc kubenswrapper[4948]: I0220 08:18:03.495808 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-h7xhj" Feb 20 08:18:03 crc kubenswrapper[4948]: W0220 08:18:03.521411 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2c199acd_2fd3_4b9f_b50c_a266191c4777.slice/crio-c448fbd8c6e26ac85c565ff365fb80b2550c88cdb6506d1a672f568a361f309d WatchSource:0}: Error finding container c448fbd8c6e26ac85c565ff365fb80b2550c88cdb6506d1a672f568a361f309d: Status 404 returned error can't find the container with id c448fbd8c6e26ac85c565ff365fb80b2550c88cdb6506d1a672f568a361f309d Feb 20 08:18:03 crc kubenswrapper[4948]: I0220 08:18:03.766274 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-c6sqk" event={"ID":"68ee0185-13d4-45c7-9a39-bb474e9314d0","Type":"ContainerStarted","Data":"05b7f34196b4aa6a1d50d3eab40559177ce6fd571f9dce95d25544ee5991268a"} Feb 20 08:18:03 crc kubenswrapper[4948]: I0220 08:18:03.766339 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-69bbfbf88f-c6sqk" Feb 20 08:18:03 crc kubenswrapper[4948]: I0220 08:18:03.767766 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-h7xhj" event={"ID":"2c199acd-2fd3-4b9f-b50c-a266191c4777","Type":"ContainerStarted","Data":"c448fbd8c6e26ac85c565ff365fb80b2550c88cdb6506d1a672f568a361f309d"} Feb 20 08:18:03 crc kubenswrapper[4948]: I0220 08:18:03.791211 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-69bbfbf88f-c6sqk" podStartSLOduration=2.791195977 podStartE2EDuration="2.791195977s" podCreationTimestamp="2026-02-20 08:18:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:18:03.787057433 +0000 UTC m=+732.761552273" watchObservedRunningTime="2026-02-20 08:18:03.791195977 +0000 UTC m=+732.765690797" Feb 20 08:18:04 crc kubenswrapper[4948]: I0220 08:18:04.778467 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-h7xhj" event={"ID":"2c199acd-2fd3-4b9f-b50c-a266191c4777","Type":"ContainerStarted","Data":"bb02cb1ad5d33d591b81172c3773b8ac4a801b8f5d3d3ab8ef4c2647b1e2845c"} Feb 20 08:18:04 crc kubenswrapper[4948]: I0220 08:18:04.779043 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-h7xhj" Feb 20 08:18:04 crc kubenswrapper[4948]: I0220 08:18:04.779066 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-h7xhj" event={"ID":"2c199acd-2fd3-4b9f-b50c-a266191c4777","Type":"ContainerStarted","Data":"8c6273d649cf80ad250c82bc059cd053d11c3997a59f0fdb4f94f58150d57ecc"} Feb 20 08:18:04 crc kubenswrapper[4948]: I0220 08:18:04.813037 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-h7xhj" podStartSLOduration=3.813019675 podStartE2EDuration="3.813019675s" podCreationTimestamp="2026-02-20 08:18:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:18:04.808662146 +0000 UTC m=+733.783156966" watchObservedRunningTime="2026-02-20 08:18:04.813019675 +0000 UTC m=+733.787514495" Feb 20 08:18:08 crc kubenswrapper[4948]: I0220 08:18:08.024958 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:18:08 crc kubenswrapper[4948]: I0220 08:18:08.025344 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:18:09 crc kubenswrapper[4948]: I0220 08:18:09.813037 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7zsxs" event={"ID":"2506d4e2-b434-4fe4-970e-7cd14601677d","Type":"ContainerStarted","Data":"3e76415ec36aa24893d0adf7e5a2c2da8d861d518e53e80711fde32279aaa0f6"} Feb 20 08:18:09 crc kubenswrapper[4948]: I0220 08:18:09.814226 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7zsxs" Feb 20 08:18:09 crc kubenswrapper[4948]: I0220 08:18:09.816248 4948 generic.go:334] "Generic (PLEG): container finished" podID="0131d05b-619d-43d8-aa38-ad6e9ce52e11" containerID="a9a7882b7d062e23fac330c2ac913b9373cfdb27dbe52941c158234dc75308b7" exitCode=0 Feb 20 08:18:09 crc kubenswrapper[4948]: I0220 08:18:09.816326 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-k66nh" event={"ID":"0131d05b-619d-43d8-aa38-ad6e9ce52e11","Type":"ContainerDied","Data":"a9a7882b7d062e23fac330c2ac913b9373cfdb27dbe52941c158234dc75308b7"} Feb 20 08:18:09 crc kubenswrapper[4948]: I0220 08:18:09.851426 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7zsxs" podStartSLOduration=2.25378656 podStartE2EDuration="8.851392599s" podCreationTimestamp="2026-02-20 08:18:01 +0000 UTC" firstStartedPulling="2026-02-20 08:18:02.406143267 +0000 UTC m=+731.380638137" lastFinishedPulling="2026-02-20 08:18:09.003749326 +0000 UTC m=+737.978244176" observedRunningTime="2026-02-20 08:18:09.842360474 +0000 UTC m=+738.816855334" watchObservedRunningTime="2026-02-20 08:18:09.851392599 +0000 UTC m=+738.825887459" Feb 20 08:18:10 crc kubenswrapper[4948]: I0220 08:18:10.824798 4948 generic.go:334] "Generic (PLEG): container finished" podID="0131d05b-619d-43d8-aa38-ad6e9ce52e11" containerID="48cfb755827c44a7b77441317598f5844ce40cd1199c0cf03cdcf7ff06e3a166" exitCode=0 Feb 20 08:18:10 crc kubenswrapper[4948]: I0220 08:18:10.824909 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-k66nh" event={"ID":"0131d05b-619d-43d8-aa38-ad6e9ce52e11","Type":"ContainerDied","Data":"48cfb755827c44a7b77441317598f5844ce40cd1199c0cf03cdcf7ff06e3a166"} Feb 20 08:18:11 crc kubenswrapper[4948]: I0220 08:18:11.833176 4948 generic.go:334] "Generic (PLEG): container finished" podID="0131d05b-619d-43d8-aa38-ad6e9ce52e11" containerID="77dbe5ef30ba0d2bfca47dad947f0aa69e25dab5fcc9c3b9cb4e9edb8e2d9e4c" exitCode=0 Feb 20 08:18:11 crc kubenswrapper[4948]: I0220 08:18:11.833281 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-k66nh" event={"ID":"0131d05b-619d-43d8-aa38-ad6e9ce52e11","Type":"ContainerDied","Data":"77dbe5ef30ba0d2bfca47dad947f0aa69e25dab5fcc9c3b9cb4e9edb8e2d9e4c"} Feb 20 08:18:12 crc kubenswrapper[4948]: I0220 08:18:12.031524 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-69bbfbf88f-c6sqk" Feb 20 08:18:12 crc kubenswrapper[4948]: I0220 08:18:12.848404 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-k66nh" event={"ID":"0131d05b-619d-43d8-aa38-ad6e9ce52e11","Type":"ContainerStarted","Data":"30cb9cbaefc7e5bc426ae70da6190592a3e466396c71864cfac194785f1b5e02"} Feb 20 08:18:12 crc kubenswrapper[4948]: I0220 08:18:12.848719 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-k66nh" event={"ID":"0131d05b-619d-43d8-aa38-ad6e9ce52e11","Type":"ContainerStarted","Data":"ed0ce7eb66e6ec5dbf50f538a5187d98f3d7beb4cdb41c88e4cb386935a9b5c3"} Feb 20 08:18:12 crc kubenswrapper[4948]: I0220 08:18:12.848736 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-k66nh" event={"ID":"0131d05b-619d-43d8-aa38-ad6e9ce52e11","Type":"ContainerStarted","Data":"a8ae56528ccc400d0e41549e4070b82a360a0f4a92a643c778435fe75b521045"} Feb 20 08:18:12 crc kubenswrapper[4948]: I0220 08:18:12.848749 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-k66nh" event={"ID":"0131d05b-619d-43d8-aa38-ad6e9ce52e11","Type":"ContainerStarted","Data":"b3b777ca08391e90927b0bb613a5a0899da7dca095fb57f79f9c3ced5d93525d"} Feb 20 08:18:12 crc kubenswrapper[4948]: I0220 08:18:12.848760 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-k66nh" event={"ID":"0131d05b-619d-43d8-aa38-ad6e9ce52e11","Type":"ContainerStarted","Data":"7fa91c580d6df65b49dc14afe2ea7940e9da4a8ffaca47686baa88ca958dab97"} Feb 20 08:18:13 crc kubenswrapper[4948]: I0220 08:18:13.501246 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-h7xhj" Feb 20 08:18:13 crc kubenswrapper[4948]: I0220 08:18:13.860280 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-k66nh" event={"ID":"0131d05b-619d-43d8-aa38-ad6e9ce52e11","Type":"ContainerStarted","Data":"5953ccf4702bfdc91094e1f472c02fc87dbf0a81b8c3ec30c9401220c436b465"} Feb 20 08:18:13 crc kubenswrapper[4948]: I0220 08:18:13.860581 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:13 crc kubenswrapper[4948]: I0220 08:18:13.897679 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-k66nh" podStartSLOduration=6.505809565 podStartE2EDuration="12.897657018s" podCreationTimestamp="2026-02-20 08:18:01 +0000 UTC" firstStartedPulling="2026-02-20 08:18:02.636767194 +0000 UTC m=+731.611262014" lastFinishedPulling="2026-02-20 08:18:09.028614617 +0000 UTC m=+738.003109467" observedRunningTime="2026-02-20 08:18:13.892175181 +0000 UTC m=+742.866670051" watchObservedRunningTime="2026-02-20 08:18:13.897657018 +0000 UTC m=+742.872151848" Feb 20 08:18:16 crc kubenswrapper[4948]: I0220 08:18:16.192119 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-r6bgs"] Feb 20 08:18:16 crc kubenswrapper[4948]: I0220 08:18:16.193728 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-r6bgs" Feb 20 08:18:16 crc kubenswrapper[4948]: I0220 08:18:16.197145 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Feb 20 08:18:16 crc kubenswrapper[4948]: I0220 08:18:16.197594 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Feb 20 08:18:16 crc kubenswrapper[4948]: I0220 08:18:16.199564 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-7bd79" Feb 20 08:18:16 crc kubenswrapper[4948]: I0220 08:18:16.223174 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-r6bgs"] Feb 20 08:18:16 crc kubenswrapper[4948]: I0220 08:18:16.248055 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrlrf\" (UniqueName: \"kubernetes.io/projected/7598c911-bd3c-4772-99f8-668c1bba4495-kube-api-access-jrlrf\") pod \"openstack-operator-index-r6bgs\" (UID: \"7598c911-bd3c-4772-99f8-668c1bba4495\") " pod="openstack-operators/openstack-operator-index-r6bgs" Feb 20 08:18:16 crc kubenswrapper[4948]: I0220 08:18:16.349505 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrlrf\" (UniqueName: \"kubernetes.io/projected/7598c911-bd3c-4772-99f8-668c1bba4495-kube-api-access-jrlrf\") pod \"openstack-operator-index-r6bgs\" (UID: \"7598c911-bd3c-4772-99f8-668c1bba4495\") " pod="openstack-operators/openstack-operator-index-r6bgs" Feb 20 08:18:16 crc kubenswrapper[4948]: I0220 08:18:16.371435 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrlrf\" (UniqueName: \"kubernetes.io/projected/7598c911-bd3c-4772-99f8-668c1bba4495-kube-api-access-jrlrf\") pod \"openstack-operator-index-r6bgs\" (UID: \"7598c911-bd3c-4772-99f8-668c1bba4495\") " pod="openstack-operators/openstack-operator-index-r6bgs" Feb 20 08:18:16 crc kubenswrapper[4948]: I0220 08:18:16.512644 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-r6bgs" Feb 20 08:18:16 crc kubenswrapper[4948]: I0220 08:18:16.754344 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-r6bgs"] Feb 20 08:18:16 crc kubenswrapper[4948]: W0220 08:18:16.765506 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7598c911_bd3c_4772_99f8_668c1bba4495.slice/crio-84871b79f61cd574bf003c73f0343d6e3cd10c676f9811ef0b6e00f040669452 WatchSource:0}: Error finding container 84871b79f61cd574bf003c73f0343d6e3cd10c676f9811ef0b6e00f040669452: Status 404 returned error can't find the container with id 84871b79f61cd574bf003c73f0343d6e3cd10c676f9811ef0b6e00f040669452 Feb 20 08:18:16 crc kubenswrapper[4948]: I0220 08:18:16.883799 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-r6bgs" event={"ID":"7598c911-bd3c-4772-99f8-668c1bba4495","Type":"ContainerStarted","Data":"84871b79f61cd574bf003c73f0343d6e3cd10c676f9811ef0b6e00f040669452"} Feb 20 08:18:17 crc kubenswrapper[4948]: I0220 08:18:17.519332 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:17 crc kubenswrapper[4948]: I0220 08:18:17.568313 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:19 crc kubenswrapper[4948]: I0220 08:18:19.370315 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-r6bgs"] Feb 20 08:18:19 crc kubenswrapper[4948]: I0220 08:18:19.905770 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-r6bgs" event={"ID":"7598c911-bd3c-4772-99f8-668c1bba4495","Type":"ContainerStarted","Data":"5bda4722a9fc484672b67eadaa90074e99796b98e5bc9ff3ced68ecd54a8925b"} Feb 20 08:18:19 crc kubenswrapper[4948]: I0220 08:18:19.906026 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-r6bgs" podUID="7598c911-bd3c-4772-99f8-668c1bba4495" containerName="registry-server" containerID="cri-o://5bda4722a9fc484672b67eadaa90074e99796b98e5bc9ff3ced68ecd54a8925b" gracePeriod=2 Feb 20 08:18:19 crc kubenswrapper[4948]: I0220 08:18:19.941622 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-r6bgs" podStartSLOduration=1.4406791239999999 podStartE2EDuration="3.941591164s" podCreationTimestamp="2026-02-20 08:18:16 +0000 UTC" firstStartedPulling="2026-02-20 08:18:16.768320043 +0000 UTC m=+745.742814873" lastFinishedPulling="2026-02-20 08:18:19.269232103 +0000 UTC m=+748.243726913" observedRunningTime="2026-02-20 08:18:19.931621275 +0000 UTC m=+748.906116155" watchObservedRunningTime="2026-02-20 08:18:19.941591164 +0000 UTC m=+748.916086024" Feb 20 08:18:19 crc kubenswrapper[4948]: I0220 08:18:19.971005 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-9z4jj"] Feb 20 08:18:19 crc kubenswrapper[4948]: I0220 08:18:19.971852 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-9z4jj" Feb 20 08:18:19 crc kubenswrapper[4948]: I0220 08:18:19.988182 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-9z4jj"] Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.026436 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrwkr\" (UniqueName: \"kubernetes.io/projected/64872326-fb95-4bc1-a6c5-0b34242883f9-kube-api-access-lrwkr\") pod \"openstack-operator-index-9z4jj\" (UID: \"64872326-fb95-4bc1-a6c5-0b34242883f9\") " pod="openstack-operators/openstack-operator-index-9z4jj" Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.127770 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrwkr\" (UniqueName: \"kubernetes.io/projected/64872326-fb95-4bc1-a6c5-0b34242883f9-kube-api-access-lrwkr\") pod \"openstack-operator-index-9z4jj\" (UID: \"64872326-fb95-4bc1-a6c5-0b34242883f9\") " pod="openstack-operators/openstack-operator-index-9z4jj" Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.166806 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrwkr\" (UniqueName: \"kubernetes.io/projected/64872326-fb95-4bc1-a6c5-0b34242883f9-kube-api-access-lrwkr\") pod \"openstack-operator-index-9z4jj\" (UID: \"64872326-fb95-4bc1-a6c5-0b34242883f9\") " pod="openstack-operators/openstack-operator-index-9z4jj" Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.283370 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-r6bgs" Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.363336 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-9z4jj" Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.449683 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrlrf\" (UniqueName: \"kubernetes.io/projected/7598c911-bd3c-4772-99f8-668c1bba4495-kube-api-access-jrlrf\") pod \"7598c911-bd3c-4772-99f8-668c1bba4495\" (UID: \"7598c911-bd3c-4772-99f8-668c1bba4495\") " Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.454133 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7598c911-bd3c-4772-99f8-668c1bba4495-kube-api-access-jrlrf" (OuterVolumeSpecName: "kube-api-access-jrlrf") pod "7598c911-bd3c-4772-99f8-668c1bba4495" (UID: "7598c911-bd3c-4772-99f8-668c1bba4495"). InnerVolumeSpecName "kube-api-access-jrlrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.553534 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrlrf\" (UniqueName: \"kubernetes.io/projected/7598c911-bd3c-4772-99f8-668c1bba4495-kube-api-access-jrlrf\") on node \"crc\" DevicePath \"\"" Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.759478 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-9z4jj"] Feb 20 08:18:20 crc kubenswrapper[4948]: W0220 08:18:20.769785 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64872326_fb95_4bc1_a6c5_0b34242883f9.slice/crio-bd0dff1cb53c89e153782945c052c068ebbad59155c54bd5132f267675f97e82 WatchSource:0}: Error finding container bd0dff1cb53c89e153782945c052c068ebbad59155c54bd5132f267675f97e82: Status 404 returned error can't find the container with id bd0dff1cb53c89e153782945c052c068ebbad59155c54bd5132f267675f97e82 Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.915366 4948 generic.go:334] "Generic (PLEG): container finished" podID="7598c911-bd3c-4772-99f8-668c1bba4495" containerID="5bda4722a9fc484672b67eadaa90074e99796b98e5bc9ff3ced68ecd54a8925b" exitCode=0 Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.915502 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-r6bgs" event={"ID":"7598c911-bd3c-4772-99f8-668c1bba4495","Type":"ContainerDied","Data":"5bda4722a9fc484672b67eadaa90074e99796b98e5bc9ff3ced68ecd54a8925b"} Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.915512 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-r6bgs" Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.915819 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-r6bgs" event={"ID":"7598c911-bd3c-4772-99f8-668c1bba4495","Type":"ContainerDied","Data":"84871b79f61cd574bf003c73f0343d6e3cd10c676f9811ef0b6e00f040669452"} Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.915862 4948 scope.go:117] "RemoveContainer" containerID="5bda4722a9fc484672b67eadaa90074e99796b98e5bc9ff3ced68ecd54a8925b" Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.917201 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-9z4jj" event={"ID":"64872326-fb95-4bc1-a6c5-0b34242883f9","Type":"ContainerStarted","Data":"bd0dff1cb53c89e153782945c052c068ebbad59155c54bd5132f267675f97e82"} Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.936347 4948 scope.go:117] "RemoveContainer" containerID="5bda4722a9fc484672b67eadaa90074e99796b98e5bc9ff3ced68ecd54a8925b" Feb 20 08:18:20 crc kubenswrapper[4948]: E0220 08:18:20.936925 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bda4722a9fc484672b67eadaa90074e99796b98e5bc9ff3ced68ecd54a8925b\": container with ID starting with 5bda4722a9fc484672b67eadaa90074e99796b98e5bc9ff3ced68ecd54a8925b not found: ID does not exist" containerID="5bda4722a9fc484672b67eadaa90074e99796b98e5bc9ff3ced68ecd54a8925b" Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.936958 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bda4722a9fc484672b67eadaa90074e99796b98e5bc9ff3ced68ecd54a8925b"} err="failed to get container status \"5bda4722a9fc484672b67eadaa90074e99796b98e5bc9ff3ced68ecd54a8925b\": rpc error: code = NotFound desc = could not find container \"5bda4722a9fc484672b67eadaa90074e99796b98e5bc9ff3ced68ecd54a8925b\": container with ID starting with 5bda4722a9fc484672b67eadaa90074e99796b98e5bc9ff3ced68ecd54a8925b not found: ID does not exist" Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.983420 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-r6bgs"] Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.988521 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-r6bgs"] Feb 20 08:18:21 crc kubenswrapper[4948]: I0220 08:18:21.733151 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7598c911-bd3c-4772-99f8-668c1bba4495" path="/var/lib/kubelet/pods/7598c911-bd3c-4772-99f8-668c1bba4495/volumes" Feb 20 08:18:21 crc kubenswrapper[4948]: I0220 08:18:21.928363 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-9z4jj" event={"ID":"64872326-fb95-4bc1-a6c5-0b34242883f9","Type":"ContainerStarted","Data":"9e917cc57712b696be945e95695645211dadf6c831fb65d4dd1e15d4de7f4f41"} Feb 20 08:18:21 crc kubenswrapper[4948]: I0220 08:18:21.945013 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7zsxs" Feb 20 08:18:21 crc kubenswrapper[4948]: I0220 08:18:21.950586 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-9z4jj" podStartSLOduration=2.896904293 podStartE2EDuration="2.950560525s" podCreationTimestamp="2026-02-20 08:18:19 +0000 UTC" firstStartedPulling="2026-02-20 08:18:20.779283199 +0000 UTC m=+749.753778049" lastFinishedPulling="2026-02-20 08:18:20.832939461 +0000 UTC m=+749.807434281" observedRunningTime="2026-02-20 08:18:21.948014641 +0000 UTC m=+750.922509491" watchObservedRunningTime="2026-02-20 08:18:21.950560525 +0000 UTC m=+750.925055385" Feb 20 08:18:22 crc kubenswrapper[4948]: I0220 08:18:22.523318 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:30 crc kubenswrapper[4948]: I0220 08:18:30.364542 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-9z4jj" Feb 20 08:18:30 crc kubenswrapper[4948]: I0220 08:18:30.365202 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-9z4jj" Feb 20 08:18:30 crc kubenswrapper[4948]: I0220 08:18:30.406163 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-9z4jj" Feb 20 08:18:31 crc kubenswrapper[4948]: I0220 08:18:31.034161 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-9z4jj" Feb 20 08:18:35 crc kubenswrapper[4948]: I0220 08:18:35.217688 4948 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 20 08:18:38 crc kubenswrapper[4948]: I0220 08:18:38.025472 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:18:38 crc kubenswrapper[4948]: I0220 08:18:38.025883 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:18:38 crc kubenswrapper[4948]: I0220 08:18:38.025942 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:18:38 crc kubenswrapper[4948]: I0220 08:18:38.026662 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b5405a03f834a339ac758e6efb1e31a8c094b3f6b973ea8540eefc716ba7bf7f"} pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 08:18:38 crc kubenswrapper[4948]: I0220 08:18:38.026729 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" containerID="cri-o://b5405a03f834a339ac758e6efb1e31a8c094b3f6b973ea8540eefc716ba7bf7f" gracePeriod=600 Feb 20 08:18:38 crc kubenswrapper[4948]: I0220 08:18:38.967366 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6"] Feb 20 08:18:38 crc kubenswrapper[4948]: E0220 08:18:38.968323 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7598c911-bd3c-4772-99f8-668c1bba4495" containerName="registry-server" Feb 20 08:18:38 crc kubenswrapper[4948]: I0220 08:18:38.968724 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="7598c911-bd3c-4772-99f8-668c1bba4495" containerName="registry-server" Feb 20 08:18:38 crc kubenswrapper[4948]: I0220 08:18:38.968890 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="7598c911-bd3c-4772-99f8-668c1bba4495" containerName="registry-server" Feb 20 08:18:38 crc kubenswrapper[4948]: I0220 08:18:38.969908 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" Feb 20 08:18:38 crc kubenswrapper[4948]: I0220 08:18:38.972124 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-z6fxt" Feb 20 08:18:38 crc kubenswrapper[4948]: I0220 08:18:38.979347 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6"] Feb 20 08:18:39 crc kubenswrapper[4948]: I0220 08:18:39.030182 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-util\") pod \"b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6\" (UID: \"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf\") " pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" Feb 20 08:18:39 crc kubenswrapper[4948]: I0220 08:18:39.030254 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tr5qj\" (UniqueName: \"kubernetes.io/projected/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-kube-api-access-tr5qj\") pod \"b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6\" (UID: \"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf\") " pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" Feb 20 08:18:39 crc kubenswrapper[4948]: I0220 08:18:39.030434 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-bundle\") pod \"b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6\" (UID: \"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf\") " pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" Feb 20 08:18:39 crc kubenswrapper[4948]: I0220 08:18:39.067387 4948 generic.go:334] "Generic (PLEG): container finished" podID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerID="b5405a03f834a339ac758e6efb1e31a8c094b3f6b973ea8540eefc716ba7bf7f" exitCode=0 Feb 20 08:18:39 crc kubenswrapper[4948]: I0220 08:18:39.067416 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerDied","Data":"b5405a03f834a339ac758e6efb1e31a8c094b3f6b973ea8540eefc716ba7bf7f"} Feb 20 08:18:39 crc kubenswrapper[4948]: I0220 08:18:39.067762 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"df8e056c06ac241c46a927c4e3f5a28f39d042bf13a8c672499d528786dbbbb0"} Feb 20 08:18:39 crc kubenswrapper[4948]: I0220 08:18:39.067859 4948 scope.go:117] "RemoveContainer" containerID="a2c247070dfb4d0488751bb46e5543e834b3853c983d98f00e419c6aabfdeb7c" Feb 20 08:18:39 crc kubenswrapper[4948]: I0220 08:18:39.131724 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-bundle\") pod \"b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6\" (UID: \"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf\") " pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" Feb 20 08:18:39 crc kubenswrapper[4948]: I0220 08:18:39.131889 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-util\") pod \"b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6\" (UID: \"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf\") " pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" Feb 20 08:18:39 crc kubenswrapper[4948]: I0220 08:18:39.132092 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tr5qj\" (UniqueName: \"kubernetes.io/projected/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-kube-api-access-tr5qj\") pod \"b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6\" (UID: \"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf\") " pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" Feb 20 08:18:39 crc kubenswrapper[4948]: I0220 08:18:39.132547 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-bundle\") pod \"b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6\" (UID: \"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf\") " pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" Feb 20 08:18:39 crc kubenswrapper[4948]: I0220 08:18:39.133060 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-util\") pod \"b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6\" (UID: \"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf\") " pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" Feb 20 08:18:39 crc kubenswrapper[4948]: I0220 08:18:39.163697 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tr5qj\" (UniqueName: \"kubernetes.io/projected/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-kube-api-access-tr5qj\") pod \"b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6\" (UID: \"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf\") " pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" Feb 20 08:18:39 crc kubenswrapper[4948]: I0220 08:18:39.291042 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" Feb 20 08:18:39 crc kubenswrapper[4948]: I0220 08:18:39.476137 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6"] Feb 20 08:18:39 crc kubenswrapper[4948]: W0220 08:18:39.489468 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda1b99eb4_2ab2_4755_92ed_a5226d91bfaf.slice/crio-56819cc9445f12d4a1429b2a438986922d555b3060e1030ae43bca1e151aca92 WatchSource:0}: Error finding container 56819cc9445f12d4a1429b2a438986922d555b3060e1030ae43bca1e151aca92: Status 404 returned error can't find the container with id 56819cc9445f12d4a1429b2a438986922d555b3060e1030ae43bca1e151aca92 Feb 20 08:18:40 crc kubenswrapper[4948]: I0220 08:18:40.076605 4948 generic.go:334] "Generic (PLEG): container finished" podID="a1b99eb4-2ab2-4755-92ed-a5226d91bfaf" containerID="33c724572b1849d1e782827fa80b211bb380e77669c568898c85b28655569808" exitCode=0 Feb 20 08:18:40 crc kubenswrapper[4948]: I0220 08:18:40.076656 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" event={"ID":"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf","Type":"ContainerDied","Data":"33c724572b1849d1e782827fa80b211bb380e77669c568898c85b28655569808"} Feb 20 08:18:40 crc kubenswrapper[4948]: I0220 08:18:40.077093 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" event={"ID":"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf","Type":"ContainerStarted","Data":"56819cc9445f12d4a1429b2a438986922d555b3060e1030ae43bca1e151aca92"} Feb 20 08:18:41 crc kubenswrapper[4948]: I0220 08:18:41.091665 4948 generic.go:334] "Generic (PLEG): container finished" podID="a1b99eb4-2ab2-4755-92ed-a5226d91bfaf" containerID="82d5bea9ab2ae66c0e5fba37f18d49d29e12c46a302a57e13775cc41280a1d46" exitCode=0 Feb 20 08:18:41 crc kubenswrapper[4948]: I0220 08:18:41.091815 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" event={"ID":"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf","Type":"ContainerDied","Data":"82d5bea9ab2ae66c0e5fba37f18d49d29e12c46a302a57e13775cc41280a1d46"} Feb 20 08:18:42 crc kubenswrapper[4948]: I0220 08:18:42.105959 4948 generic.go:334] "Generic (PLEG): container finished" podID="a1b99eb4-2ab2-4755-92ed-a5226d91bfaf" containerID="0bac1a873808460bd3bbd795e40f16ef73fcda97d242286fa6c592d23ce18f0e" exitCode=0 Feb 20 08:18:42 crc kubenswrapper[4948]: I0220 08:18:42.106088 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" event={"ID":"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf","Type":"ContainerDied","Data":"0bac1a873808460bd3bbd795e40f16ef73fcda97d242286fa6c592d23ce18f0e"} Feb 20 08:18:43 crc kubenswrapper[4948]: I0220 08:18:43.508560 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" Feb 20 08:18:43 crc kubenswrapper[4948]: I0220 08:18:43.609519 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-util\") pod \"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf\" (UID: \"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf\") " Feb 20 08:18:43 crc kubenswrapper[4948]: I0220 08:18:43.609930 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-bundle\") pod \"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf\" (UID: \"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf\") " Feb 20 08:18:43 crc kubenswrapper[4948]: I0220 08:18:43.609974 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tr5qj\" (UniqueName: \"kubernetes.io/projected/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-kube-api-access-tr5qj\") pod \"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf\" (UID: \"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf\") " Feb 20 08:18:43 crc kubenswrapper[4948]: I0220 08:18:43.610633 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-bundle" (OuterVolumeSpecName: "bundle") pod "a1b99eb4-2ab2-4755-92ed-a5226d91bfaf" (UID: "a1b99eb4-2ab2-4755-92ed-a5226d91bfaf"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:18:43 crc kubenswrapper[4948]: I0220 08:18:43.616680 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-kube-api-access-tr5qj" (OuterVolumeSpecName: "kube-api-access-tr5qj") pod "a1b99eb4-2ab2-4755-92ed-a5226d91bfaf" (UID: "a1b99eb4-2ab2-4755-92ed-a5226d91bfaf"). InnerVolumeSpecName "kube-api-access-tr5qj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:18:43 crc kubenswrapper[4948]: I0220 08:18:43.629167 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-util" (OuterVolumeSpecName: "util") pod "a1b99eb4-2ab2-4755-92ed-a5226d91bfaf" (UID: "a1b99eb4-2ab2-4755-92ed-a5226d91bfaf"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:18:43 crc kubenswrapper[4948]: I0220 08:18:43.712145 4948 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-util\") on node \"crc\" DevicePath \"\"" Feb 20 08:18:43 crc kubenswrapper[4948]: I0220 08:18:43.712198 4948 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:18:43 crc kubenswrapper[4948]: I0220 08:18:43.712219 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tr5qj\" (UniqueName: \"kubernetes.io/projected/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-kube-api-access-tr5qj\") on node \"crc\" DevicePath \"\"" Feb 20 08:18:44 crc kubenswrapper[4948]: I0220 08:18:44.127020 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" event={"ID":"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf","Type":"ContainerDied","Data":"56819cc9445f12d4a1429b2a438986922d555b3060e1030ae43bca1e151aca92"} Feb 20 08:18:44 crc kubenswrapper[4948]: I0220 08:18:44.127079 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="56819cc9445f12d4a1429b2a438986922d555b3060e1030ae43bca1e151aca92" Feb 20 08:18:44 crc kubenswrapper[4948]: I0220 08:18:44.127115 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" Feb 20 08:18:51 crc kubenswrapper[4948]: I0220 08:18:51.239199 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-59c446d8d6-4cs2n"] Feb 20 08:18:51 crc kubenswrapper[4948]: E0220 08:18:51.240111 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1b99eb4-2ab2-4755-92ed-a5226d91bfaf" containerName="util" Feb 20 08:18:51 crc kubenswrapper[4948]: I0220 08:18:51.240128 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1b99eb4-2ab2-4755-92ed-a5226d91bfaf" containerName="util" Feb 20 08:18:51 crc kubenswrapper[4948]: E0220 08:18:51.240150 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1b99eb4-2ab2-4755-92ed-a5226d91bfaf" containerName="extract" Feb 20 08:18:51 crc kubenswrapper[4948]: I0220 08:18:51.240157 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1b99eb4-2ab2-4755-92ed-a5226d91bfaf" containerName="extract" Feb 20 08:18:51 crc kubenswrapper[4948]: E0220 08:18:51.240172 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1b99eb4-2ab2-4755-92ed-a5226d91bfaf" containerName="pull" Feb 20 08:18:51 crc kubenswrapper[4948]: I0220 08:18:51.240196 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1b99eb4-2ab2-4755-92ed-a5226d91bfaf" containerName="pull" Feb 20 08:18:51 crc kubenswrapper[4948]: I0220 08:18:51.240321 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1b99eb4-2ab2-4755-92ed-a5226d91bfaf" containerName="extract" Feb 20 08:18:51 crc kubenswrapper[4948]: I0220 08:18:51.240735 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-59c446d8d6-4cs2n" Feb 20 08:18:51 crc kubenswrapper[4948]: I0220 08:18:51.243771 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-9hmd7" Feb 20 08:18:51 crc kubenswrapper[4948]: I0220 08:18:51.323234 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-59c446d8d6-4cs2n"] Feb 20 08:18:51 crc kubenswrapper[4948]: I0220 08:18:51.323780 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4899m\" (UniqueName: \"kubernetes.io/projected/9465a7a8-c08e-44c9-a76d-3878b0bff3af-kube-api-access-4899m\") pod \"openstack-operator-controller-init-59c446d8d6-4cs2n\" (UID: \"9465a7a8-c08e-44c9-a76d-3878b0bff3af\") " pod="openstack-operators/openstack-operator-controller-init-59c446d8d6-4cs2n" Feb 20 08:18:51 crc kubenswrapper[4948]: I0220 08:18:51.424667 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4899m\" (UniqueName: \"kubernetes.io/projected/9465a7a8-c08e-44c9-a76d-3878b0bff3af-kube-api-access-4899m\") pod \"openstack-operator-controller-init-59c446d8d6-4cs2n\" (UID: \"9465a7a8-c08e-44c9-a76d-3878b0bff3af\") " pod="openstack-operators/openstack-operator-controller-init-59c446d8d6-4cs2n" Feb 20 08:18:51 crc kubenswrapper[4948]: I0220 08:18:51.444817 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4899m\" (UniqueName: \"kubernetes.io/projected/9465a7a8-c08e-44c9-a76d-3878b0bff3af-kube-api-access-4899m\") pod \"openstack-operator-controller-init-59c446d8d6-4cs2n\" (UID: \"9465a7a8-c08e-44c9-a76d-3878b0bff3af\") " pod="openstack-operators/openstack-operator-controller-init-59c446d8d6-4cs2n" Feb 20 08:18:51 crc kubenswrapper[4948]: I0220 08:18:51.574710 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-59c446d8d6-4cs2n" Feb 20 08:18:51 crc kubenswrapper[4948]: I0220 08:18:51.789039 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-59c446d8d6-4cs2n"] Feb 20 08:18:51 crc kubenswrapper[4948]: W0220 08:18:51.798733 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9465a7a8_c08e_44c9_a76d_3878b0bff3af.slice/crio-151160100b0cd6df4f324891edcdc7d3493e3f9a61b069450e63c6482f0a924c WatchSource:0}: Error finding container 151160100b0cd6df4f324891edcdc7d3493e3f9a61b069450e63c6482f0a924c: Status 404 returned error can't find the container with id 151160100b0cd6df4f324891edcdc7d3493e3f9a61b069450e63c6482f0a924c Feb 20 08:18:52 crc kubenswrapper[4948]: I0220 08:18:52.207891 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-59c446d8d6-4cs2n" event={"ID":"9465a7a8-c08e-44c9-a76d-3878b0bff3af","Type":"ContainerStarted","Data":"151160100b0cd6df4f324891edcdc7d3493e3f9a61b069450e63c6482f0a924c"} Feb 20 08:18:56 crc kubenswrapper[4948]: I0220 08:18:56.247370 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-59c446d8d6-4cs2n" event={"ID":"9465a7a8-c08e-44c9-a76d-3878b0bff3af","Type":"ContainerStarted","Data":"a4b82b549f81b88e36ef7dc8eb38618f1e90051ff837cadf4dd63fa70291dd3f"} Feb 20 08:18:56 crc kubenswrapper[4948]: I0220 08:18:56.247869 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-59c446d8d6-4cs2n" Feb 20 08:18:56 crc kubenswrapper[4948]: I0220 08:18:56.279504 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-59c446d8d6-4cs2n" podStartSLOduration=1.715627904 podStartE2EDuration="5.279480194s" podCreationTimestamp="2026-02-20 08:18:51 +0000 UTC" firstStartedPulling="2026-02-20 08:18:51.800480279 +0000 UTC m=+780.774975109" lastFinishedPulling="2026-02-20 08:18:55.364332569 +0000 UTC m=+784.338827399" observedRunningTime="2026-02-20 08:18:56.279167687 +0000 UTC m=+785.253662517" watchObservedRunningTime="2026-02-20 08:18:56.279480194 +0000 UTC m=+785.253975054" Feb 20 08:19:01 crc kubenswrapper[4948]: I0220 08:19:01.577668 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-59c446d8d6-4cs2n" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.536276 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-4h9v5"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.537791 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-4h9v5" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.540146 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-5fbgl" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.543379 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5d946d989d-c6xqd"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.544430 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-c6xqd" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.546235 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-vspgb" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.550991 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-4h9v5"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.557499 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5d946d989d-c6xqd"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.566070 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-kjpkr"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.567401 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-kjpkr" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.575625 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-r8cz9" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.613927 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-kjpkr"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.631068 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-6cr5p"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.631790 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-6cr5p" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.633191 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-qt8sb" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.642752 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987464f4-xc74s"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.643702 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987464f4-xc74s" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.650541 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zgxs\" (UniqueName: \"kubernetes.io/projected/586113f2-38b6-4bd1-8adb-3c155bb35ba9-kube-api-access-7zgxs\") pod \"cinder-operator-controller-manager-5d946d989d-c6xqd\" (UID: \"586113f2-38b6-4bd1-8adb-3c155bb35ba9\") " pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-c6xqd" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.650603 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwh5b\" (UniqueName: \"kubernetes.io/projected/b8e12d0c-9564-4bf2-ac61-b22d2fbdf855-kube-api-access-gwh5b\") pod \"barbican-operator-controller-manager-868647ff47-4h9v5\" (UID: \"b8e12d0c-9564-4bf2-ac61-b22d2fbdf855\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-4h9v5" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.650615 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-6m82q" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.650635 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9zrw\" (UniqueName: \"kubernetes.io/projected/98a1aa06-948b-4034-bc07-7e546e341a8f-kube-api-access-d9zrw\") pod \"designate-operator-controller-manager-6d8bf5c495-kjpkr\" (UID: \"98a1aa06-948b-4034-bc07-7e546e341a8f\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-kjpkr" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.651039 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987464f4-xc74s"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.657753 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-6cr5p"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.659815 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-bkl49"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.660894 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-bkl49" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.673586 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-57g8d" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.679093 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-bkl49"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.682072 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.682769 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.684642 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.687183 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-nxfl4" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.721681 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-5fqzw"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.722606 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-5fqzw" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.729417 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-5vbfv" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.742183 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.742224 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-5fqzw"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.753765 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gr4jf\" (UniqueName: \"kubernetes.io/projected/2d303d1a-1062-4bf6-be68-1bd6d1a3228f-kube-api-access-gr4jf\") pod \"heat-operator-controller-manager-69f49c598c-6cr5p\" (UID: \"2d303d1a-1062-4bf6-be68-1bd6d1a3228f\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-6cr5p" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.753820 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rd7r\" (UniqueName: \"kubernetes.io/projected/4a663826-d4b5-4ed3-8270-099b003390b4-kube-api-access-8rd7r\") pod \"horizon-operator-controller-manager-5b9b8895d5-bkl49\" (UID: \"4a663826-d4b5-4ed3-8270-099b003390b4\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-bkl49" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.753892 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ht8br\" (UniqueName: \"kubernetes.io/projected/347cdd7c-e5e8-49b1-a4f1-687b8a06b250-kube-api-access-ht8br\") pod \"glance-operator-controller-manager-77987464f4-xc74s\" (UID: \"347cdd7c-e5e8-49b1-a4f1-687b8a06b250\") " pod="openstack-operators/glance-operator-controller-manager-77987464f4-xc74s" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.753944 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zgxs\" (UniqueName: \"kubernetes.io/projected/586113f2-38b6-4bd1-8adb-3c155bb35ba9-kube-api-access-7zgxs\") pod \"cinder-operator-controller-manager-5d946d989d-c6xqd\" (UID: \"586113f2-38b6-4bd1-8adb-3c155bb35ba9\") " pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-c6xqd" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.754009 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwh5b\" (UniqueName: \"kubernetes.io/projected/b8e12d0c-9564-4bf2-ac61-b22d2fbdf855-kube-api-access-gwh5b\") pod \"barbican-operator-controller-manager-868647ff47-4h9v5\" (UID: \"b8e12d0c-9564-4bf2-ac61-b22d2fbdf855\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-4h9v5" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.754042 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9zrw\" (UniqueName: \"kubernetes.io/projected/98a1aa06-948b-4034-bc07-7e546e341a8f-kube-api-access-d9zrw\") pod \"designate-operator-controller-manager-6d8bf5c495-kjpkr\" (UID: \"98a1aa06-948b-4034-bc07-7e546e341a8f\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-kjpkr" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.756741 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-54f6768c69-lj2sz"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.757781 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-lj2sz" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.759640 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-b44t8" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.765651 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-ttkfm"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.766650 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-ttkfm" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.771832 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-qhs6f" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.792178 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9zrw\" (UniqueName: \"kubernetes.io/projected/98a1aa06-948b-4034-bc07-7e546e341a8f-kube-api-access-d9zrw\") pod \"designate-operator-controller-manager-6d8bf5c495-kjpkr\" (UID: \"98a1aa06-948b-4034-bc07-7e546e341a8f\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-kjpkr" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.793334 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-54f6768c69-lj2sz"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.800715 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwh5b\" (UniqueName: \"kubernetes.io/projected/b8e12d0c-9564-4bf2-ac61-b22d2fbdf855-kube-api-access-gwh5b\") pod \"barbican-operator-controller-manager-868647ff47-4h9v5\" (UID: \"b8e12d0c-9564-4bf2-ac61-b22d2fbdf855\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-4h9v5" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.805158 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zgxs\" (UniqueName: \"kubernetes.io/projected/586113f2-38b6-4bd1-8adb-3c155bb35ba9-kube-api-access-7zgxs\") pod \"cinder-operator-controller-manager-5d946d989d-c6xqd\" (UID: \"586113f2-38b6-4bd1-8adb-3c155bb35ba9\") " pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-c6xqd" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.808882 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-zcccd"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.809784 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-zcccd" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.814872 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-2njl7" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.817363 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-ttkfm"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.856574 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ht8br\" (UniqueName: \"kubernetes.io/projected/347cdd7c-e5e8-49b1-a4f1-687b8a06b250-kube-api-access-ht8br\") pod \"glance-operator-controller-manager-77987464f4-xc74s\" (UID: \"347cdd7c-e5e8-49b1-a4f1-687b8a06b250\") " pod="openstack-operators/glance-operator-controller-manager-77987464f4-xc74s" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.856651 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbh26\" (UniqueName: \"kubernetes.io/projected/93b93501-acfe-4274-9a22-ca644b1d11d3-kube-api-access-mbh26\") pod \"infra-operator-controller-manager-79d975b745-jrlw5\" (UID: \"93b93501-acfe-4274-9a22-ca644b1d11d3\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.856681 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvjkg\" (UniqueName: \"kubernetes.io/projected/105a118c-121d-4582-960f-1da9957980cb-kube-api-access-hvjkg\") pod \"ironic-operator-controller-manager-554564d7fc-5fqzw\" (UID: \"105a118c-121d-4582-960f-1da9957980cb\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-5fqzw" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.856741 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert\") pod \"infra-operator-controller-manager-79d975b745-jrlw5\" (UID: \"93b93501-acfe-4274-9a22-ca644b1d11d3\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.856779 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gr4jf\" (UniqueName: \"kubernetes.io/projected/2d303d1a-1062-4bf6-be68-1bd6d1a3228f-kube-api-access-gr4jf\") pod \"heat-operator-controller-manager-69f49c598c-6cr5p\" (UID: \"2d303d1a-1062-4bf6-be68-1bd6d1a3228f\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-6cr5p" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.856801 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rd7r\" (UniqueName: \"kubernetes.io/projected/4a663826-d4b5-4ed3-8270-099b003390b4-kube-api-access-8rd7r\") pod \"horizon-operator-controller-manager-5b9b8895d5-bkl49\" (UID: \"4a663826-d4b5-4ed3-8270-099b003390b4\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-bkl49" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.856897 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28rlg\" (UniqueName: \"kubernetes.io/projected/3fbbd11d-612d-479a-a34f-505d995a4871-kube-api-access-28rlg\") pod \"keystone-operator-controller-manager-b4d948c87-ttkfm\" (UID: \"3fbbd11d-612d-479a-a34f-505d995a4871\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-ttkfm" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.856984 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pphn\" (UniqueName: \"kubernetes.io/projected/f762ec0d-e09c-4baf-9540-f4ac61ce7234-kube-api-access-6pphn\") pod \"manila-operator-controller-manager-54f6768c69-lj2sz\" (UID: \"f762ec0d-e09c-4baf-9540-f4ac61ce7234\") " pod="openstack-operators/manila-operator-controller-manager-54f6768c69-lj2sz" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.868021 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-4h9v5" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.876079 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64ddbf8bb-jcnxd"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.879456 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-jcnxd" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.880439 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ht8br\" (UniqueName: \"kubernetes.io/projected/347cdd7c-e5e8-49b1-a4f1-687b8a06b250-kube-api-access-ht8br\") pod \"glance-operator-controller-manager-77987464f4-xc74s\" (UID: \"347cdd7c-e5e8-49b1-a4f1-687b8a06b250\") " pod="openstack-operators/glance-operator-controller-manager-77987464f4-xc74s" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.880723 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rd7r\" (UniqueName: \"kubernetes.io/projected/4a663826-d4b5-4ed3-8270-099b003390b4-kube-api-access-8rd7r\") pod \"horizon-operator-controller-manager-5b9b8895d5-bkl49\" (UID: \"4a663826-d4b5-4ed3-8270-099b003390b4\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-bkl49" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.880835 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gr4jf\" (UniqueName: \"kubernetes.io/projected/2d303d1a-1062-4bf6-be68-1bd6d1a3228f-kube-api-access-gr4jf\") pod \"heat-operator-controller-manager-69f49c598c-6cr5p\" (UID: \"2d303d1a-1062-4bf6-be68-1bd6d1a3228f\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-6cr5p" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.882217 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-v67vx" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.882592 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-c6xqd" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.890464 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-jxwsx"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.891396 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-jxwsx" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.893107 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-72llz" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.896508 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-zcccd"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.897391 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-kjpkr" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.903932 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64ddbf8bb-jcnxd"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.910225 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f8888797-hl5w7"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.911772 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-hl5w7" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.918320 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-q8jj9" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.929093 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-jxwsx"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.933980 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f8888797-hl5w7"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.940572 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.942071 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.944048 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.944420 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-c7nf5" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.953448 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-d44cf6b75-dkkq9"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.954355 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-dkkq9" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.956958 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-nhpm5" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.957630 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-znrkz"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.957680 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert\") pod \"infra-operator-controller-manager-79d975b745-jrlw5\" (UID: \"93b93501-acfe-4274-9a22-ca644b1d11d3\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.957727 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28rlg\" (UniqueName: \"kubernetes.io/projected/3fbbd11d-612d-479a-a34f-505d995a4871-kube-api-access-28rlg\") pod \"keystone-operator-controller-manager-b4d948c87-ttkfm\" (UID: \"3fbbd11d-612d-479a-a34f-505d995a4871\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-ttkfm" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.957746 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pphn\" (UniqueName: \"kubernetes.io/projected/f762ec0d-e09c-4baf-9540-f4ac61ce7234-kube-api-access-6pphn\") pod \"manila-operator-controller-manager-54f6768c69-lj2sz\" (UID: \"f762ec0d-e09c-4baf-9540-f4ac61ce7234\") " pod="openstack-operators/manila-operator-controller-manager-54f6768c69-lj2sz" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.957773 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjkn5\" (UniqueName: \"kubernetes.io/projected/6e8e8b87-5a00-486f-a00c-1450c109f3b2-kube-api-access-vjkn5\") pod \"mariadb-operator-controller-manager-6994f66f48-zcccd\" (UID: \"6e8e8b87-5a00-486f-a00c-1450c109f3b2\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-zcccd" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.957809 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbh26\" (UniqueName: \"kubernetes.io/projected/93b93501-acfe-4274-9a22-ca644b1d11d3-kube-api-access-mbh26\") pod \"infra-operator-controller-manager-79d975b745-jrlw5\" (UID: \"93b93501-acfe-4274-9a22-ca644b1d11d3\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.957835 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvjkg\" (UniqueName: \"kubernetes.io/projected/105a118c-121d-4582-960f-1da9957980cb-kube-api-access-hvjkg\") pod \"ironic-operator-controller-manager-554564d7fc-5fqzw\" (UID: \"105a118c-121d-4582-960f-1da9957980cb\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-5fqzw" Feb 20 08:19:39 crc kubenswrapper[4948]: E0220 08:19:39.958105 4948 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 20 08:19:39 crc kubenswrapper[4948]: E0220 08:19:39.958154 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert podName:93b93501-acfe-4274-9a22-ca644b1d11d3 nodeName:}" failed. No retries permitted until 2026-02-20 08:19:40.458137939 +0000 UTC m=+829.432632759 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert") pod "infra-operator-controller-manager-79d975b745-jrlw5" (UID: "93b93501-acfe-4274-9a22-ca644b1d11d3") : secret "infra-operator-webhook-server-cert" not found Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.958262 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-znrkz" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.964619 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-vvqgl" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.966314 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-6cr5p" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.981316 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987464f4-xc74s" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.985684 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvjkg\" (UniqueName: \"kubernetes.io/projected/105a118c-121d-4582-960f-1da9957980cb-kube-api-access-hvjkg\") pod \"ironic-operator-controller-manager-554564d7fc-5fqzw\" (UID: \"105a118c-121d-4582-960f-1da9957980cb\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-5fqzw" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.985835 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbh26\" (UniqueName: \"kubernetes.io/projected/93b93501-acfe-4274-9a22-ca644b1d11d3-kube-api-access-mbh26\") pod \"infra-operator-controller-manager-79d975b745-jrlw5\" (UID: \"93b93501-acfe-4274-9a22-ca644b1d11d3\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.991651 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28rlg\" (UniqueName: \"kubernetes.io/projected/3fbbd11d-612d-479a-a34f-505d995a4871-kube-api-access-28rlg\") pod \"keystone-operator-controller-manager-b4d948c87-ttkfm\" (UID: \"3fbbd11d-612d-479a-a34f-505d995a4871\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-ttkfm" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.995537 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pphn\" (UniqueName: \"kubernetes.io/projected/f762ec0d-e09c-4baf-9540-f4ac61ce7234-kube-api-access-6pphn\") pod \"manila-operator-controller-manager-54f6768c69-lj2sz\" (UID: \"f762ec0d-e09c-4baf-9540-f4ac61ce7234\") " pod="openstack-operators/manila-operator-controller-manager-54f6768c69-lj2sz" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.007128 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-bkl49" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.027632 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-d44cf6b75-dkkq9"] Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.027672 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-znrkz"] Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.035153 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m"] Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.045458 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-5lfts"] Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.048508 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-68f46476f-5lfts" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.054469 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-wmrt2" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.054672 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-5lfts"] Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.054946 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-5fqzw" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.059006 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6x5rb\" (UniqueName: \"kubernetes.io/projected/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-kube-api-access-6x5rb\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m\" (UID: \"b8b07685-095c-4bbd-a30d-57ca59d7cbdc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.059060 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgpnv\" (UniqueName: \"kubernetes.io/projected/ee33bd02-a3ce-415d-9d2d-fefd383e9810-kube-api-access-fgpnv\") pod \"octavia-operator-controller-manager-69f8888797-hl5w7\" (UID: \"ee33bd02-a3ce-415d-9d2d-fefd383e9810\") " pod="openstack-operators/octavia-operator-controller-manager-69f8888797-hl5w7" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.059106 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xw4bv\" (UniqueName: \"kubernetes.io/projected/88142137-864d-4660-a688-a7dcc503851b-kube-api-access-xw4bv\") pod \"neutron-operator-controller-manager-64ddbf8bb-jcnxd\" (UID: \"88142137-864d-4660-a688-a7dcc503851b\") " pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-jcnxd" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.059138 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9srrq\" (UniqueName: \"kubernetes.io/projected/16933747-642c-45ff-9f98-9321c633826a-kube-api-access-9srrq\") pod \"placement-operator-controller-manager-8497b45c89-znrkz\" (UID: \"16933747-642c-45ff-9f98-9321c633826a\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-znrkz" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.059162 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kd2fq\" (UniqueName: \"kubernetes.io/projected/7c7bb531-7900-4cc6-9d9b-bae52dabc59d-kube-api-access-kd2fq\") pod \"nova-operator-controller-manager-567668f5cf-jxwsx\" (UID: \"7c7bb531-7900-4cc6-9d9b-bae52dabc59d\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-jxwsx" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.059221 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m\" (UID: \"b8b07685-095c-4bbd-a30d-57ca59d7cbdc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.059255 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2csx\" (UniqueName: \"kubernetes.io/projected/3ea675de-d1b0-4880-9652-eb066f6b0fb7-kube-api-access-q2csx\") pod \"ovn-operator-controller-manager-d44cf6b75-dkkq9\" (UID: \"3ea675de-d1b0-4880-9652-eb066f6b0fb7\") " pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-dkkq9" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.059311 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjkn5\" (UniqueName: \"kubernetes.io/projected/6e8e8b87-5a00-486f-a00c-1450c109f3b2-kube-api-access-vjkn5\") pod \"mariadb-operator-controller-manager-6994f66f48-zcccd\" (UID: \"6e8e8b87-5a00-486f-a00c-1450c109f3b2\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-zcccd" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.083394 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-lj2sz" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.084721 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjkn5\" (UniqueName: \"kubernetes.io/projected/6e8e8b87-5a00-486f-a00c-1450c109f3b2-kube-api-access-vjkn5\") pod \"mariadb-operator-controller-manager-6994f66f48-zcccd\" (UID: \"6e8e8b87-5a00-486f-a00c-1450c109f3b2\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-zcccd" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.091289 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-5rrd9"] Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.093713 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-5rrd9" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.095547 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-5rrd9"] Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.095760 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-8psrm" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.154339 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-ttkfm" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.156990 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-7866795846-f7cxn"] Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.157842 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-7866795846-f7cxn" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.160674 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m\" (UID: \"b8b07685-095c-4bbd-a30d-57ca59d7cbdc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.160718 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2csx\" (UniqueName: \"kubernetes.io/projected/3ea675de-d1b0-4880-9652-eb066f6b0fb7-kube-api-access-q2csx\") pod \"ovn-operator-controller-manager-d44cf6b75-dkkq9\" (UID: \"3ea675de-d1b0-4880-9652-eb066f6b0fb7\") " pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-dkkq9" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.160777 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxmkg\" (UniqueName: \"kubernetes.io/projected/d65e0993-11a0-4e81-963f-eeb9dcb92536-kube-api-access-xxmkg\") pod \"swift-operator-controller-manager-68f46476f-5lfts\" (UID: \"d65e0993-11a0-4e81-963f-eeb9dcb92536\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-5lfts" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.160805 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6x5rb\" (UniqueName: \"kubernetes.io/projected/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-kube-api-access-6x5rb\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m\" (UID: \"b8b07685-095c-4bbd-a30d-57ca59d7cbdc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.160831 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgpnv\" (UniqueName: \"kubernetes.io/projected/ee33bd02-a3ce-415d-9d2d-fefd383e9810-kube-api-access-fgpnv\") pod \"octavia-operator-controller-manager-69f8888797-hl5w7\" (UID: \"ee33bd02-a3ce-415d-9d2d-fefd383e9810\") " pod="openstack-operators/octavia-operator-controller-manager-69f8888797-hl5w7" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.160851 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xw4bv\" (UniqueName: \"kubernetes.io/projected/88142137-864d-4660-a688-a7dcc503851b-kube-api-access-xw4bv\") pod \"neutron-operator-controller-manager-64ddbf8bb-jcnxd\" (UID: \"88142137-864d-4660-a688-a7dcc503851b\") " pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-jcnxd" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.160871 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9srrq\" (UniqueName: \"kubernetes.io/projected/16933747-642c-45ff-9f98-9321c633826a-kube-api-access-9srrq\") pod \"placement-operator-controller-manager-8497b45c89-znrkz\" (UID: \"16933747-642c-45ff-9f98-9321c633826a\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-znrkz" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.160889 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kd2fq\" (UniqueName: \"kubernetes.io/projected/7c7bb531-7900-4cc6-9d9b-bae52dabc59d-kube-api-access-kd2fq\") pod \"nova-operator-controller-manager-567668f5cf-jxwsx\" (UID: \"7c7bb531-7900-4cc6-9d9b-bae52dabc59d\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-jxwsx" Feb 20 08:19:40 crc kubenswrapper[4948]: E0220 08:19:40.161266 4948 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 08:19:40 crc kubenswrapper[4948]: E0220 08:19:40.161313 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert podName:b8b07685-095c-4bbd-a30d-57ca59d7cbdc nodeName:}" failed. No retries permitted until 2026-02-20 08:19:40.66129961 +0000 UTC m=+829.635794430 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" (UID: "b8b07685-095c-4bbd-a30d-57ca59d7cbdc") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.161375 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-6zx75" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.176309 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-zcccd" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.179455 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kd2fq\" (UniqueName: \"kubernetes.io/projected/7c7bb531-7900-4cc6-9d9b-bae52dabc59d-kube-api-access-kd2fq\") pod \"nova-operator-controller-manager-567668f5cf-jxwsx\" (UID: \"7c7bb531-7900-4cc6-9d9b-bae52dabc59d\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-jxwsx" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.182540 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2csx\" (UniqueName: \"kubernetes.io/projected/3ea675de-d1b0-4880-9652-eb066f6b0fb7-kube-api-access-q2csx\") pod \"ovn-operator-controller-manager-d44cf6b75-dkkq9\" (UID: \"3ea675de-d1b0-4880-9652-eb066f6b0fb7\") " pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-dkkq9" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.186959 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9srrq\" (UniqueName: \"kubernetes.io/projected/16933747-642c-45ff-9f98-9321c633826a-kube-api-access-9srrq\") pod \"placement-operator-controller-manager-8497b45c89-znrkz\" (UID: \"16933747-642c-45ff-9f98-9321c633826a\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-znrkz" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.188745 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgpnv\" (UniqueName: \"kubernetes.io/projected/ee33bd02-a3ce-415d-9d2d-fefd383e9810-kube-api-access-fgpnv\") pod \"octavia-operator-controller-manager-69f8888797-hl5w7\" (UID: \"ee33bd02-a3ce-415d-9d2d-fefd383e9810\") " pod="openstack-operators/octavia-operator-controller-manager-69f8888797-hl5w7" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.188833 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-7866795846-f7cxn"] Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.197171 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xw4bv\" (UniqueName: \"kubernetes.io/projected/88142137-864d-4660-a688-a7dcc503851b-kube-api-access-xw4bv\") pod \"neutron-operator-controller-manager-64ddbf8bb-jcnxd\" (UID: \"88142137-864d-4660-a688-a7dcc503851b\") " pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-jcnxd" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.201747 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6x5rb\" (UniqueName: \"kubernetes.io/projected/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-kube-api-access-6x5rb\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m\" (UID: \"b8b07685-095c-4bbd-a30d-57ca59d7cbdc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.237755 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5db88f68c-pndq7"] Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.238960 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-pndq7" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.242882 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-qsgww" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.251071 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5db88f68c-pndq7"] Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.259614 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-jcnxd" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.261448 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxmkg\" (UniqueName: \"kubernetes.io/projected/d65e0993-11a0-4e81-963f-eeb9dcb92536-kube-api-access-xxmkg\") pod \"swift-operator-controller-manager-68f46476f-5lfts\" (UID: \"d65e0993-11a0-4e81-963f-eeb9dcb92536\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-5lfts" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.261482 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rjvk\" (UniqueName: \"kubernetes.io/projected/0f98bc58-d0e3-405b-88fd-d8bd65f415a4-kube-api-access-8rjvk\") pod \"test-operator-controller-manager-7866795846-f7cxn\" (UID: \"0f98bc58-d0e3-405b-88fd-d8bd65f415a4\") " pod="openstack-operators/test-operator-controller-manager-7866795846-f7cxn" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.261511 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2f9x\" (UniqueName: \"kubernetes.io/projected/0f0cbb65-3c8e-41e6-8059-34e121de0821-kube-api-access-w2f9x\") pod \"telemetry-operator-controller-manager-7f45b4ff68-5rrd9\" (UID: \"0f0cbb65-3c8e-41e6-8059-34e121de0821\") " pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-5rrd9" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.283063 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxmkg\" (UniqueName: \"kubernetes.io/projected/d65e0993-11a0-4e81-963f-eeb9dcb92536-kube-api-access-xxmkg\") pod \"swift-operator-controller-manager-68f46476f-5lfts\" (UID: \"d65e0993-11a0-4e81-963f-eeb9dcb92536\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-5lfts" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.284308 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-jxwsx" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.286747 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk"] Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.288394 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.291300 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.292184 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.293861 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-5sbbv" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.295195 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk"] Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.302982 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-hl5w7" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.335891 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-59pcs"] Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.341569 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-59pcs" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.355680 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-9rn66" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.362906 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rjvk\" (UniqueName: \"kubernetes.io/projected/0f98bc58-d0e3-405b-88fd-d8bd65f415a4-kube-api-access-8rjvk\") pod \"test-operator-controller-manager-7866795846-f7cxn\" (UID: \"0f98bc58-d0e3-405b-88fd-d8bd65f415a4\") " pod="openstack-operators/test-operator-controller-manager-7866795846-f7cxn" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.364092 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scvsh\" (UniqueName: \"kubernetes.io/projected/f2400173-aa87-476c-8216-4f8c9cf9d474-kube-api-access-scvsh\") pod \"watcher-operator-controller-manager-5db88f68c-pndq7\" (UID: \"f2400173-aa87-476c-8216-4f8c9cf9d474\") " pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-pndq7" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.364218 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2f9x\" (UniqueName: \"kubernetes.io/projected/0f0cbb65-3c8e-41e6-8059-34e121de0821-kube-api-access-w2f9x\") pod \"telemetry-operator-controller-manager-7f45b4ff68-5rrd9\" (UID: \"0f0cbb65-3c8e-41e6-8059-34e121de0821\") " pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-5rrd9" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.386786 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2f9x\" (UniqueName: \"kubernetes.io/projected/0f0cbb65-3c8e-41e6-8059-34e121de0821-kube-api-access-w2f9x\") pod \"telemetry-operator-controller-manager-7f45b4ff68-5rrd9\" (UID: \"0f0cbb65-3c8e-41e6-8059-34e121de0821\") " pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-5rrd9" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.386920 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rjvk\" (UniqueName: \"kubernetes.io/projected/0f98bc58-d0e3-405b-88fd-d8bd65f415a4-kube-api-access-8rjvk\") pod \"test-operator-controller-manager-7866795846-f7cxn\" (UID: \"0f98bc58-d0e3-405b-88fd-d8bd65f415a4\") " pod="openstack-operators/test-operator-controller-manager-7866795846-f7cxn" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.396241 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-59pcs"] Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.397285 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-dkkq9" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.420379 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-znrkz" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.427701 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5d946d989d-c6xqd"] Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.451307 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-68f46476f-5lfts" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.470722 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scvsh\" (UniqueName: \"kubernetes.io/projected/f2400173-aa87-476c-8216-4f8c9cf9d474-kube-api-access-scvsh\") pod \"watcher-operator-controller-manager-5db88f68c-pndq7\" (UID: \"f2400173-aa87-476c-8216-4f8c9cf9d474\") " pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-pndq7" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.470785 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqz2k\" (UniqueName: \"kubernetes.io/projected/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-kube-api-access-pqz2k\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.470832 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert\") pod \"infra-operator-controller-manager-79d975b745-jrlw5\" (UID: \"93b93501-acfe-4274-9a22-ca644b1d11d3\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.470850 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5r6fn\" (UniqueName: \"kubernetes.io/projected/fe0aca33-1acf-463a-91a6-bbf35a38fd7c-kube-api-access-5r6fn\") pod \"rabbitmq-cluster-operator-manager-668c99d594-59pcs\" (UID: \"fe0aca33-1acf-463a-91a6-bbf35a38fd7c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-59pcs" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.470891 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.470916 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:40 crc kubenswrapper[4948]: E0220 08:19:40.471097 4948 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 20 08:19:40 crc kubenswrapper[4948]: E0220 08:19:40.471191 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert podName:93b93501-acfe-4274-9a22-ca644b1d11d3 nodeName:}" failed. No retries permitted until 2026-02-20 08:19:41.471165879 +0000 UTC m=+830.445660769 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert") pod "infra-operator-controller-manager-79d975b745-jrlw5" (UID: "93b93501-acfe-4274-9a22-ca644b1d11d3") : secret "infra-operator-webhook-server-cert" not found Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.490251 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scvsh\" (UniqueName: \"kubernetes.io/projected/f2400173-aa87-476c-8216-4f8c9cf9d474-kube-api-access-scvsh\") pod \"watcher-operator-controller-manager-5db88f68c-pndq7\" (UID: \"f2400173-aa87-476c-8216-4f8c9cf9d474\") " pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-pndq7" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.500629 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-5rrd9" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.555592 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-7866795846-f7cxn" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.565675 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-pndq7" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.572795 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqz2k\" (UniqueName: \"kubernetes.io/projected/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-kube-api-access-pqz2k\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.572881 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5r6fn\" (UniqueName: \"kubernetes.io/projected/fe0aca33-1acf-463a-91a6-bbf35a38fd7c-kube-api-access-5r6fn\") pod \"rabbitmq-cluster-operator-manager-668c99d594-59pcs\" (UID: \"fe0aca33-1acf-463a-91a6-bbf35a38fd7c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-59pcs" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.572930 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.572963 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:40 crc kubenswrapper[4948]: E0220 08:19:40.573111 4948 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 20 08:19:40 crc kubenswrapper[4948]: E0220 08:19:40.573176 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs podName:e9f13bcd-ac5a-4cbd-952a-a5b9dceee562 nodeName:}" failed. No retries permitted until 2026-02-20 08:19:41.073158109 +0000 UTC m=+830.047652929 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs") pod "openstack-operator-controller-manager-8569bc6fdb-lt4jk" (UID: "e9f13bcd-ac5a-4cbd-952a-a5b9dceee562") : secret "webhook-server-cert" not found Feb 20 08:19:40 crc kubenswrapper[4948]: E0220 08:19:40.573230 4948 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 20 08:19:40 crc kubenswrapper[4948]: E0220 08:19:40.573255 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs podName:e9f13bcd-ac5a-4cbd-952a-a5b9dceee562 nodeName:}" failed. No retries permitted until 2026-02-20 08:19:41.073245821 +0000 UTC m=+830.047740651 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs") pod "openstack-operator-controller-manager-8569bc6fdb-lt4jk" (UID: "e9f13bcd-ac5a-4cbd-952a-a5b9dceee562") : secret "metrics-server-cert" not found Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.592539 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5r6fn\" (UniqueName: \"kubernetes.io/projected/fe0aca33-1acf-463a-91a6-bbf35a38fd7c-kube-api-access-5r6fn\") pod \"rabbitmq-cluster-operator-manager-668c99d594-59pcs\" (UID: \"fe0aca33-1acf-463a-91a6-bbf35a38fd7c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-59pcs" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.594345 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqz2k\" (UniqueName: \"kubernetes.io/projected/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-kube-api-access-pqz2k\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.242494 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-59pcs" Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.255356 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m\" (UID: \"b8b07685-095c-4bbd-a30d-57ca59d7cbdc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.255441 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.255505 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:41 crc kubenswrapper[4948]: E0220 08:19:41.255615 4948 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 08:19:41 crc kubenswrapper[4948]: E0220 08:19:41.255687 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert podName:b8b07685-095c-4bbd-a30d-57ca59d7cbdc nodeName:}" failed. No retries permitted until 2026-02-20 08:19:42.255666653 +0000 UTC m=+831.230161473 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" (UID: "b8b07685-095c-4bbd-a30d-57ca59d7cbdc") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 08:19:41 crc kubenswrapper[4948]: E0220 08:19:41.255802 4948 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 20 08:19:41 crc kubenswrapper[4948]: E0220 08:19:41.255866 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs podName:e9f13bcd-ac5a-4cbd-952a-a5b9dceee562 nodeName:}" failed. No retries permitted until 2026-02-20 08:19:42.255843947 +0000 UTC m=+831.230338837 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs") pod "openstack-operator-controller-manager-8569bc6fdb-lt4jk" (UID: "e9f13bcd-ac5a-4cbd-952a-a5b9dceee562") : secret "webhook-server-cert" not found Feb 20 08:19:41 crc kubenswrapper[4948]: E0220 08:19:41.255941 4948 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 20 08:19:41 crc kubenswrapper[4948]: E0220 08:19:41.256025 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs podName:e9f13bcd-ac5a-4cbd-952a-a5b9dceee562 nodeName:}" failed. No retries permitted until 2026-02-20 08:19:42.256007081 +0000 UTC m=+831.230501901 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs") pod "openstack-operator-controller-manager-8569bc6fdb-lt4jk" (UID: "e9f13bcd-ac5a-4cbd-952a-a5b9dceee562") : secret "metrics-server-cert" not found Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.292022 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-kjpkr"] Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.345492 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-c6xqd" event={"ID":"586113f2-38b6-4bd1-8adb-3c155bb35ba9","Type":"ContainerStarted","Data":"13047a876b037c4521e3c794420c31f09e64c07bb3c2a460d3e03b1558f19e0a"} Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.366280 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987464f4-xc74s"] Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.386398 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-4h9v5"] Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.394788 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-6cr5p"] Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.574172 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert\") pod \"infra-operator-controller-manager-79d975b745-jrlw5\" (UID: \"93b93501-acfe-4274-9a22-ca644b1d11d3\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" Feb 20 08:19:41 crc kubenswrapper[4948]: E0220 08:19:41.574389 4948 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 20 08:19:41 crc kubenswrapper[4948]: E0220 08:19:41.574454 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert podName:93b93501-acfe-4274-9a22-ca644b1d11d3 nodeName:}" failed. No retries permitted until 2026-02-20 08:19:43.574420467 +0000 UTC m=+832.548915287 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert") pod "infra-operator-controller-manager-79d975b745-jrlw5" (UID: "93b93501-acfe-4274-9a22-ca644b1d11d3") : secret "infra-operator-webhook-server-cert" not found Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.658209 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-zcccd"] Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.749478 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-jxwsx"] Feb 20 08:19:41 crc kubenswrapper[4948]: W0220 08:19:41.750470 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4a663826_d4b5_4ed3_8270_099b003390b4.slice/crio-d5056dc2edcf92c09e73d1ddb7aadce15676bf58b6f724db6c57780f6d5d150e WatchSource:0}: Error finding container d5056dc2edcf92c09e73d1ddb7aadce15676bf58b6f724db6c57780f6d5d150e: Status 404 returned error can't find the container with id d5056dc2edcf92c09e73d1ddb7aadce15676bf58b6f724db6c57780f6d5d150e Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.752997 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-bkl49"] Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.780529 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-54f6768c69-lj2sz"] Feb 20 08:19:41 crc kubenswrapper[4948]: W0220 08:19:41.787073 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf762ec0d_e09c_4baf_9540_f4ac61ce7234.slice/crio-70b50a2a70093089e8c102ecbbaf96cb245f8d0ea9b5229b84fb487d1b462c43 WatchSource:0}: Error finding container 70b50a2a70093089e8c102ecbbaf96cb245f8d0ea9b5229b84fb487d1b462c43: Status 404 returned error can't find the container with id 70b50a2a70093089e8c102ecbbaf96cb245f8d0ea9b5229b84fb487d1b462c43 Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.803566 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64ddbf8bb-jcnxd"] Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.824604 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f8888797-hl5w7"] Feb 20 08:19:41 crc kubenswrapper[4948]: W0220 08:19:41.827168 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod105a118c_121d_4582_960f_1da9957980cb.slice/crio-00e9662c8d49f6ee5c0d2f5b929b7f67c58779ae298ca81d10507f0502ec2bde WatchSource:0}: Error finding container 00e9662c8d49f6ee5c0d2f5b929b7f67c58779ae298ca81d10507f0502ec2bde: Status 404 returned error can't find the container with id 00e9662c8d49f6ee5c0d2f5b929b7f67c58779ae298ca81d10507f0502ec2bde Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.842581 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-ttkfm"] Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.855842 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-5fqzw"] Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.975068 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-znrkz"] Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.985668 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-5rrd9"] Feb 20 08:19:41 crc kubenswrapper[4948]: E0220 08:19:41.990737 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:a57336b9f95b703f80453db87e43a2834ca1bdc89480796d28ebbe0a9702ecfd,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9srrq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-8497b45c89-znrkz_openstack-operators(16933747-642c-45ff-9f98-9321c633826a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 20 08:19:41 crc kubenswrapper[4948]: W0220 08:19:41.991822 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0f0cbb65_3c8e_41e6_8059_34e121de0821.slice/crio-dab83244371f9d620457757e4fa587e6437cac3702cb742b7ab71532b934f994 WatchSource:0}: Error finding container dab83244371f9d620457757e4fa587e6437cac3702cb742b7ab71532b934f994: Status 404 returned error can't find the container with id dab83244371f9d620457757e4fa587e6437cac3702cb742b7ab71532b934f994 Feb 20 08:19:41 crc kubenswrapper[4948]: E0220 08:19:41.991855 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-znrkz" podUID="16933747-642c-45ff-9f98-9321c633826a" Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.994762 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-5lfts"] Feb 20 08:19:41 crc kubenswrapper[4948]: W0220 08:19:41.995591 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3ea675de_d1b0_4880_9652_eb066f6b0fb7.slice/crio-5652c94b03f402e891c3173cd21873fba5ae10d25a358e87372237ec7407dbdf WatchSource:0}: Error finding container 5652c94b03f402e891c3173cd21873fba5ae10d25a358e87372237ec7407dbdf: Status 404 returned error can't find the container with id 5652c94b03f402e891c3173cd21873fba5ae10d25a358e87372237ec7407dbdf Feb 20 08:19:41 crc kubenswrapper[4948]: E0220 08:19:41.995634 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:66a4b9322ebb573313178ea88e31026d4532f461592b9fae2dff71efd9256d99,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-w2f9x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-7f45b4ff68-5rrd9_openstack-operators(0f0cbb65-3c8e-41e6-8059-34e121de0821): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 20 08:19:41 crc kubenswrapper[4948]: E0220 08:19:41.996945 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-5rrd9" podUID="0f0cbb65-3c8e-41e6-8059-34e121de0821" Feb 20 08:19:41 crc kubenswrapper[4948]: W0220 08:19:41.997793 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd65e0993_11a0_4e81_963f_eeb9dcb92536.slice/crio-29ab656c5b13d0755f9f86b4e797b7aeee91cc1d92ef44299791121d7aa453db WatchSource:0}: Error finding container 29ab656c5b13d0755f9f86b4e797b7aeee91cc1d92ef44299791121d7aa453db: Status 404 returned error can't find the container with id 29ab656c5b13d0755f9f86b4e797b7aeee91cc1d92ef44299791121d7aa453db Feb 20 08:19:41 crc kubenswrapper[4948]: E0220 08:19:41.998669 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:543c103838f3e6ef48755665a7695dfa3ed84753c557560257d265db31f92759,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-q2csx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-d44cf6b75-dkkq9_openstack-operators(3ea675de-d1b0-4880-9652-eb066f6b0fb7): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 20 08:19:41 crc kubenswrapper[4948]: E0220 08:19:41.999904 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-dkkq9" podUID="3ea675de-d1b0-4880-9652-eb066f6b0fb7" Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.003401 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-d44cf6b75-dkkq9"] Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.004746 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:3d676f1281e24ef07de617570d2f7fbf625032e41866d1551a856c052248bb04,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xxmkg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-68f46476f-5lfts_openstack-operators(d65e0993-11a0-4e81-963f-eeb9dcb92536): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.006110 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-68f46476f-5lfts" podUID="d65e0993-11a0-4e81-963f-eeb9dcb92536" Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.102230 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-7866795846-f7cxn"] Feb 20 08:19:42 crc kubenswrapper[4948]: W0220 08:19:42.103071 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfe0aca33_1acf_463a_91a6_bbf35a38fd7c.slice/crio-262d00936c737b141e11d8eb909d105b51ee27f7a50c4cdca6ed42fd7d060bdb WatchSource:0}: Error finding container 262d00936c737b141e11d8eb909d105b51ee27f7a50c4cdca6ed42fd7d060bdb: Status 404 returned error can't find the container with id 262d00936c737b141e11d8eb909d105b51ee27f7a50c4cdca6ed42fd7d060bdb Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.112294 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-59pcs"] Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.128954 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:f0fabdf79095def0f8b1c0442925548a94ca94bed4de2d3b171277129f8079e6,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8rjvk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-7866795846-f7cxn_openstack-operators(0f98bc58-d0e3-405b-88fd-d8bd65f415a4): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.130692 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-7866795846-f7cxn" podUID="0f98bc58-d0e3-405b-88fd-d8bd65f415a4" Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.133691 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5db88f68c-pndq7"] Feb 20 08:19:42 crc kubenswrapper[4948]: W0220 08:19:42.134965 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf2400173_aa87_476c_8216_4f8c9cf9d474.slice/crio-838f9667f35a9044f976260a4ff261c3e9a746ca96302b07edae7d2a2025dcde WatchSource:0}: Error finding container 838f9667f35a9044f976260a4ff261c3e9a746ca96302b07edae7d2a2025dcde: Status 404 returned error can't find the container with id 838f9667f35a9044f976260a4ff261c3e9a746ca96302b07edae7d2a2025dcde Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.137940 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:d01ae848290e880c09127d5297418dea40fc7f090fdab9bf2c578c7e7f53aec0,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-scvsh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-5db88f68c-pndq7_openstack-operators(f2400173-aa87-476c-8216-4f8c9cf9d474): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.139277 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-pndq7" podUID="f2400173-aa87-476c-8216-4f8c9cf9d474" Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.286680 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m\" (UID: \"b8b07685-095c-4bbd-a30d-57ca59d7cbdc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.286751 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.286792 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.286822 4948 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.286880 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert podName:b8b07685-095c-4bbd-a30d-57ca59d7cbdc nodeName:}" failed. No retries permitted until 2026-02-20 08:19:44.286862282 +0000 UTC m=+833.261357102 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" (UID: "b8b07685-095c-4bbd-a30d-57ca59d7cbdc") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.286921 4948 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.286943 4948 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.286999 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs podName:e9f13bcd-ac5a-4cbd-952a-a5b9dceee562 nodeName:}" failed. No retries permitted until 2026-02-20 08:19:44.286963744 +0000 UTC m=+833.261458614 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs") pod "openstack-operator-controller-manager-8569bc6fdb-lt4jk" (UID: "e9f13bcd-ac5a-4cbd-952a-a5b9dceee562") : secret "metrics-server-cert" not found Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.287019 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs podName:e9f13bcd-ac5a-4cbd-952a-a5b9dceee562 nodeName:}" failed. No retries permitted until 2026-02-20 08:19:44.287010645 +0000 UTC m=+833.261505585 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs") pod "openstack-operator-controller-manager-8569bc6fdb-lt4jk" (UID: "e9f13bcd-ac5a-4cbd-952a-a5b9dceee562") : secret "webhook-server-cert" not found Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.355238 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-zcccd" event={"ID":"6e8e8b87-5a00-486f-a00c-1450c109f3b2","Type":"ContainerStarted","Data":"47ae6bb8643020f78c8d6b028710c10d052d0f5f885d0e6e30a064e4ee2b1808"} Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.359598 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987464f4-xc74s" event={"ID":"347cdd7c-e5e8-49b1-a4f1-687b8a06b250","Type":"ContainerStarted","Data":"0fe0a9ff36a6df94c7f4c9e05fac924488dd011d114ed7a793ed8b30842d850b"} Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.360751 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-ttkfm" event={"ID":"3fbbd11d-612d-479a-a34f-505d995a4871","Type":"ContainerStarted","Data":"41cd793edb1558eca392ff334a39ce44122dc7272667675fbbdc34fe045cc85e"} Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.363496 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-6cr5p" event={"ID":"2d303d1a-1062-4bf6-be68-1bd6d1a3228f","Type":"ContainerStarted","Data":"58fd24496419c67ef3c86821ce7c78a64791538c2f5f99035d3770c0df3befa6"} Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.365055 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-znrkz" event={"ID":"16933747-642c-45ff-9f98-9321c633826a","Type":"ContainerStarted","Data":"b3c67bfc9b9225d9c7af76548873d10834a152e407c56ac266aea1339f7a7815"} Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.368041 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a57336b9f95b703f80453db87e43a2834ca1bdc89480796d28ebbe0a9702ecfd\\\"\"" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-znrkz" podUID="16933747-642c-45ff-9f98-9321c633826a" Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.369319 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-lj2sz" event={"ID":"f762ec0d-e09c-4baf-9540-f4ac61ce7234","Type":"ContainerStarted","Data":"70b50a2a70093089e8c102ecbbaf96cb245f8d0ea9b5229b84fb487d1b462c43"} Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.370813 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-7866795846-f7cxn" event={"ID":"0f98bc58-d0e3-405b-88fd-d8bd65f415a4","Type":"ContainerStarted","Data":"341ab804e809f94e8c08d8bf607b29a7ead898ee88d5194de5e3200b72af43e1"} Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.372345 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f0fabdf79095def0f8b1c0442925548a94ca94bed4de2d3b171277129f8079e6\\\"\"" pod="openstack-operators/test-operator-controller-manager-7866795846-f7cxn" podUID="0f98bc58-d0e3-405b-88fd-d8bd65f415a4" Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.374199 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-bkl49" event={"ID":"4a663826-d4b5-4ed3-8270-099b003390b4","Type":"ContainerStarted","Data":"d5056dc2edcf92c09e73d1ddb7aadce15676bf58b6f724db6c57780f6d5d150e"} Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.375610 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-4h9v5" event={"ID":"b8e12d0c-9564-4bf2-ac61-b22d2fbdf855","Type":"ContainerStarted","Data":"6dce06766402c643cb1b6a34926fbe50490c103bdf9617d63acffc20fa7f79d4"} Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.393645 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-59pcs" event={"ID":"fe0aca33-1acf-463a-91a6-bbf35a38fd7c","Type":"ContainerStarted","Data":"262d00936c737b141e11d8eb909d105b51ee27f7a50c4cdca6ed42fd7d060bdb"} Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.412907 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-pndq7" event={"ID":"f2400173-aa87-476c-8216-4f8c9cf9d474","Type":"ContainerStarted","Data":"838f9667f35a9044f976260a4ff261c3e9a746ca96302b07edae7d2a2025dcde"} Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.421311 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:d01ae848290e880c09127d5297418dea40fc7f090fdab9bf2c578c7e7f53aec0\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-pndq7" podUID="f2400173-aa87-476c-8216-4f8c9cf9d474" Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.423590 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-jcnxd" event={"ID":"88142137-864d-4660-a688-a7dcc503851b","Type":"ContainerStarted","Data":"fa97191b7ad9caddbfadeaee1e4083a3f35f42d90318475a6ba08698c2cc7564"} Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.425776 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-jxwsx" event={"ID":"7c7bb531-7900-4cc6-9d9b-bae52dabc59d","Type":"ContainerStarted","Data":"69fa85ce2c62a538feaade8eead3cb5277451a5c84d034cc4d6cfdc81efb8299"} Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.428018 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-kjpkr" event={"ID":"98a1aa06-948b-4034-bc07-7e546e341a8f","Type":"ContainerStarted","Data":"e0b64b507f7813caac03a0d0569c4d3882b783f1e9f4537b9c134c856f7f94a2"} Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.429215 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-68f46476f-5lfts" event={"ID":"d65e0993-11a0-4e81-963f-eeb9dcb92536","Type":"ContainerStarted","Data":"29ab656c5b13d0755f9f86b4e797b7aeee91cc1d92ef44299791121d7aa453db"} Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.432426 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3d676f1281e24ef07de617570d2f7fbf625032e41866d1551a856c052248bb04\\\"\"" pod="openstack-operators/swift-operator-controller-manager-68f46476f-5lfts" podUID="d65e0993-11a0-4e81-963f-eeb9dcb92536" Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.435566 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-hl5w7" event={"ID":"ee33bd02-a3ce-415d-9d2d-fefd383e9810","Type":"ContainerStarted","Data":"cda502a65aa7cd9591f58265268b0fd9086950692f98e8eb2fd0370fb70d2a92"} Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.443014 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-5fqzw" event={"ID":"105a118c-121d-4582-960f-1da9957980cb","Type":"ContainerStarted","Data":"00e9662c8d49f6ee5c0d2f5b929b7f67c58779ae298ca81d10507f0502ec2bde"} Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.445011 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-dkkq9" event={"ID":"3ea675de-d1b0-4880-9652-eb066f6b0fb7","Type":"ContainerStarted","Data":"5652c94b03f402e891c3173cd21873fba5ae10d25a358e87372237ec7407dbdf"} Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.450467 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:543c103838f3e6ef48755665a7695dfa3ed84753c557560257d265db31f92759\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-dkkq9" podUID="3ea675de-d1b0-4880-9652-eb066f6b0fb7" Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.450486 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-5rrd9" event={"ID":"0f0cbb65-3c8e-41e6-8059-34e121de0821","Type":"ContainerStarted","Data":"dab83244371f9d620457757e4fa587e6437cac3702cb742b7ab71532b934f994"} Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.456155 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:66a4b9322ebb573313178ea88e31026d4532f461592b9fae2dff71efd9256d99\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-5rrd9" podUID="0f0cbb65-3c8e-41e6-8059-34e121de0821" Feb 20 08:19:43 crc kubenswrapper[4948]: E0220 08:19:43.463379 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f0fabdf79095def0f8b1c0442925548a94ca94bed4de2d3b171277129f8079e6\\\"\"" pod="openstack-operators/test-operator-controller-manager-7866795846-f7cxn" podUID="0f98bc58-d0e3-405b-88fd-d8bd65f415a4" Feb 20 08:19:43 crc kubenswrapper[4948]: E0220 08:19:43.463668 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a57336b9f95b703f80453db87e43a2834ca1bdc89480796d28ebbe0a9702ecfd\\\"\"" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-znrkz" podUID="16933747-642c-45ff-9f98-9321c633826a" Feb 20 08:19:43 crc kubenswrapper[4948]: E0220 08:19:43.463426 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3d676f1281e24ef07de617570d2f7fbf625032e41866d1551a856c052248bb04\\\"\"" pod="openstack-operators/swift-operator-controller-manager-68f46476f-5lfts" podUID="d65e0993-11a0-4e81-963f-eeb9dcb92536" Feb 20 08:19:43 crc kubenswrapper[4948]: E0220 08:19:43.463739 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:66a4b9322ebb573313178ea88e31026d4532f461592b9fae2dff71efd9256d99\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-5rrd9" podUID="0f0cbb65-3c8e-41e6-8059-34e121de0821" Feb 20 08:19:43 crc kubenswrapper[4948]: E0220 08:19:43.464334 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:543c103838f3e6ef48755665a7695dfa3ed84753c557560257d265db31f92759\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-dkkq9" podUID="3ea675de-d1b0-4880-9652-eb066f6b0fb7" Feb 20 08:19:43 crc kubenswrapper[4948]: E0220 08:19:43.470718 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:d01ae848290e880c09127d5297418dea40fc7f090fdab9bf2c578c7e7f53aec0\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-pndq7" podUID="f2400173-aa87-476c-8216-4f8c9cf9d474" Feb 20 08:19:43 crc kubenswrapper[4948]: I0220 08:19:43.606352 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert\") pod \"infra-operator-controller-manager-79d975b745-jrlw5\" (UID: \"93b93501-acfe-4274-9a22-ca644b1d11d3\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" Feb 20 08:19:43 crc kubenswrapper[4948]: E0220 08:19:43.606565 4948 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 20 08:19:43 crc kubenswrapper[4948]: E0220 08:19:43.606659 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert podName:93b93501-acfe-4274-9a22-ca644b1d11d3 nodeName:}" failed. No retries permitted until 2026-02-20 08:19:47.606633469 +0000 UTC m=+836.581128309 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert") pod "infra-operator-controller-manager-79d975b745-jrlw5" (UID: "93b93501-acfe-4274-9a22-ca644b1d11d3") : secret "infra-operator-webhook-server-cert" not found Feb 20 08:19:44 crc kubenswrapper[4948]: I0220 08:19:44.313961 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m\" (UID: \"b8b07685-095c-4bbd-a30d-57ca59d7cbdc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" Feb 20 08:19:44 crc kubenswrapper[4948]: I0220 08:19:44.314256 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:44 crc kubenswrapper[4948]: I0220 08:19:44.314285 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:44 crc kubenswrapper[4948]: E0220 08:19:44.314121 4948 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 08:19:44 crc kubenswrapper[4948]: E0220 08:19:44.314473 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert podName:b8b07685-095c-4bbd-a30d-57ca59d7cbdc nodeName:}" failed. No retries permitted until 2026-02-20 08:19:48.314460158 +0000 UTC m=+837.288954968 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" (UID: "b8b07685-095c-4bbd-a30d-57ca59d7cbdc") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 08:19:44 crc kubenswrapper[4948]: E0220 08:19:44.314918 4948 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 20 08:19:44 crc kubenswrapper[4948]: E0220 08:19:44.314949 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs podName:e9f13bcd-ac5a-4cbd-952a-a5b9dceee562 nodeName:}" failed. No retries permitted until 2026-02-20 08:19:48.314940919 +0000 UTC m=+837.289435739 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs") pod "openstack-operator-controller-manager-8569bc6fdb-lt4jk" (UID: "e9f13bcd-ac5a-4cbd-952a-a5b9dceee562") : secret "metrics-server-cert" not found Feb 20 08:19:44 crc kubenswrapper[4948]: E0220 08:19:44.314424 4948 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 20 08:19:44 crc kubenswrapper[4948]: E0220 08:19:44.315029 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs podName:e9f13bcd-ac5a-4cbd-952a-a5b9dceee562 nodeName:}" failed. No retries permitted until 2026-02-20 08:19:48.315022691 +0000 UTC m=+837.289517511 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs") pod "openstack-operator-controller-manager-8569bc6fdb-lt4jk" (UID: "e9f13bcd-ac5a-4cbd-952a-a5b9dceee562") : secret "webhook-server-cert" not found Feb 20 08:19:47 crc kubenswrapper[4948]: I0220 08:19:47.663965 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert\") pod \"infra-operator-controller-manager-79d975b745-jrlw5\" (UID: \"93b93501-acfe-4274-9a22-ca644b1d11d3\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" Feb 20 08:19:47 crc kubenswrapper[4948]: E0220 08:19:47.664114 4948 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 20 08:19:47 crc kubenswrapper[4948]: E0220 08:19:47.664195 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert podName:93b93501-acfe-4274-9a22-ca644b1d11d3 nodeName:}" failed. No retries permitted until 2026-02-20 08:19:55.664177664 +0000 UTC m=+844.638672484 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert") pod "infra-operator-controller-manager-79d975b745-jrlw5" (UID: "93b93501-acfe-4274-9a22-ca644b1d11d3") : secret "infra-operator-webhook-server-cert" not found Feb 20 08:19:48 crc kubenswrapper[4948]: I0220 08:19:48.375046 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m\" (UID: \"b8b07685-095c-4bbd-a30d-57ca59d7cbdc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" Feb 20 08:19:48 crc kubenswrapper[4948]: I0220 08:19:48.375546 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:48 crc kubenswrapper[4948]: E0220 08:19:48.375252 4948 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 08:19:48 crc kubenswrapper[4948]: E0220 08:19:48.375686 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert podName:b8b07685-095c-4bbd-a30d-57ca59d7cbdc nodeName:}" failed. No retries permitted until 2026-02-20 08:19:56.375661786 +0000 UTC m=+845.350156626 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" (UID: "b8b07685-095c-4bbd-a30d-57ca59d7cbdc") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 08:19:48 crc kubenswrapper[4948]: I0220 08:19:48.375601 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:48 crc kubenswrapper[4948]: E0220 08:19:48.375729 4948 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 20 08:19:48 crc kubenswrapper[4948]: E0220 08:19:48.375744 4948 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 20 08:19:48 crc kubenswrapper[4948]: E0220 08:19:48.375813 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs podName:e9f13bcd-ac5a-4cbd-952a-a5b9dceee562 nodeName:}" failed. No retries permitted until 2026-02-20 08:19:56.375787459 +0000 UTC m=+845.350282359 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs") pod "openstack-operator-controller-manager-8569bc6fdb-lt4jk" (UID: "e9f13bcd-ac5a-4cbd-952a-a5b9dceee562") : secret "webhook-server-cert" not found Feb 20 08:19:48 crc kubenswrapper[4948]: E0220 08:19:48.375839 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs podName:e9f13bcd-ac5a-4cbd-952a-a5b9dceee562 nodeName:}" failed. No retries permitted until 2026-02-20 08:19:56.37582756 +0000 UTC m=+845.350322510 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs") pod "openstack-operator-controller-manager-8569bc6fdb-lt4jk" (UID: "e9f13bcd-ac5a-4cbd-952a-a5b9dceee562") : secret "metrics-server-cert" not found Feb 20 08:19:55 crc kubenswrapper[4948]: E0220 08:19:55.579139 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:e4689246ae78635dc3c1db9c677d8b16b8f94276df15fb9c84bfc57cc6578fcf" Feb 20 08:19:55 crc kubenswrapper[4948]: E0220 08:19:55.579757 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:e4689246ae78635dc3c1db9c677d8b16b8f94276df15fb9c84bfc57cc6578fcf,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xw4bv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-64ddbf8bb-jcnxd_openstack-operators(88142137-864d-4660-a688-a7dcc503851b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 08:19:55 crc kubenswrapper[4948]: E0220 08:19:55.581068 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-jcnxd" podUID="88142137-864d-4660-a688-a7dcc503851b" Feb 20 08:19:55 crc kubenswrapper[4948]: I0220 08:19:55.690175 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert\") pod \"infra-operator-controller-manager-79d975b745-jrlw5\" (UID: \"93b93501-acfe-4274-9a22-ca644b1d11d3\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" Feb 20 08:19:55 crc kubenswrapper[4948]: E0220 08:19:55.690398 4948 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 20 08:19:55 crc kubenswrapper[4948]: E0220 08:19:55.690449 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert podName:93b93501-acfe-4274-9a22-ca644b1d11d3 nodeName:}" failed. No retries permitted until 2026-02-20 08:20:11.690431536 +0000 UTC m=+860.664926356 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert") pod "infra-operator-controller-manager-79d975b745-jrlw5" (UID: "93b93501-acfe-4274-9a22-ca644b1d11d3") : secret "infra-operator-webhook-server-cert" not found Feb 20 08:19:56 crc kubenswrapper[4948]: I0220 08:19:56.402401 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m\" (UID: \"b8b07685-095c-4bbd-a30d-57ca59d7cbdc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" Feb 20 08:19:56 crc kubenswrapper[4948]: I0220 08:19:56.402515 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:56 crc kubenswrapper[4948]: I0220 08:19:56.402614 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:56 crc kubenswrapper[4948]: E0220 08:19:56.402731 4948 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 20 08:19:56 crc kubenswrapper[4948]: E0220 08:19:56.402826 4948 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 20 08:19:56 crc kubenswrapper[4948]: E0220 08:19:56.402856 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs podName:e9f13bcd-ac5a-4cbd-952a-a5b9dceee562 nodeName:}" failed. No retries permitted until 2026-02-20 08:20:12.40282645 +0000 UTC m=+861.377321310 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs") pod "openstack-operator-controller-manager-8569bc6fdb-lt4jk" (UID: "e9f13bcd-ac5a-4cbd-952a-a5b9dceee562") : secret "metrics-server-cert" not found Feb 20 08:19:56 crc kubenswrapper[4948]: E0220 08:19:56.403014 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs podName:e9f13bcd-ac5a-4cbd-952a-a5b9dceee562 nodeName:}" failed. No retries permitted until 2026-02-20 08:20:12.402945743 +0000 UTC m=+861.377440603 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs") pod "openstack-operator-controller-manager-8569bc6fdb-lt4jk" (UID: "e9f13bcd-ac5a-4cbd-952a-a5b9dceee562") : secret "webhook-server-cert" not found Feb 20 08:19:56 crc kubenswrapper[4948]: E0220 08:19:56.403117 4948 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 08:19:56 crc kubenswrapper[4948]: E0220 08:19:56.403226 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert podName:b8b07685-095c-4bbd-a30d-57ca59d7cbdc nodeName:}" failed. No retries permitted until 2026-02-20 08:20:12.403197189 +0000 UTC m=+861.377692099 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" (UID: "b8b07685-095c-4bbd-a30d-57ca59d7cbdc") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 08:19:56 crc kubenswrapper[4948]: E0220 08:19:56.578024 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:e4689246ae78635dc3c1db9c677d8b16b8f94276df15fb9c84bfc57cc6578fcf\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-jcnxd" podUID="88142137-864d-4660-a688-a7dcc503851b" Feb 20 08:19:56 crc kubenswrapper[4948]: E0220 08:19:56.642846 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:229fc8c8d94dd4102d2151cd4ec1eaaa09d897c2b396d06e903f61ea29c1fa34" Feb 20 08:19:56 crc kubenswrapper[4948]: E0220 08:19:56.643326 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:229fc8c8d94dd4102d2151cd4ec1eaaa09d897c2b396d06e903f61ea29c1fa34,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fgpnv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-69f8888797-hl5w7_openstack-operators(ee33bd02-a3ce-415d-9d2d-fefd383e9810): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 08:19:56 crc kubenswrapper[4948]: E0220 08:19:56.644658 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-hl5w7" podUID="ee33bd02-a3ce-415d-9d2d-fefd383e9810" Feb 20 08:19:57 crc kubenswrapper[4948]: E0220 08:19:57.348789 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:7e1b0b7b172ad0d707ab80dd72d609e1d0f5bbd38a22c24a28ed0f17a960c867" Feb 20 08:19:57 crc kubenswrapper[4948]: E0220 08:19:57.348998 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:7e1b0b7b172ad0d707ab80dd72d609e1d0f5bbd38a22c24a28ed0f17a960c867,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hvjkg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-554564d7fc-5fqzw_openstack-operators(105a118c-121d-4582-960f-1da9957980cb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 08:19:57 crc kubenswrapper[4948]: E0220 08:19:57.350260 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-5fqzw" podUID="105a118c-121d-4582-960f-1da9957980cb" Feb 20 08:19:57 crc kubenswrapper[4948]: E0220 08:19:57.583617 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:7e1b0b7b172ad0d707ab80dd72d609e1d0f5bbd38a22c24a28ed0f17a960c867\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-5fqzw" podUID="105a118c-121d-4582-960f-1da9957980cb" Feb 20 08:19:57 crc kubenswrapper[4948]: E0220 08:19:57.584781 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:229fc8c8d94dd4102d2151cd4ec1eaaa09d897c2b396d06e903f61ea29c1fa34\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-hl5w7" podUID="ee33bd02-a3ce-415d-9d2d-fefd383e9810" Feb 20 08:19:58 crc kubenswrapper[4948]: E0220 08:19:58.828394 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:fe85dd595906fac0fe1e7a42215bb306a963cf87d55e07cd2573726b690b2838" Feb 20 08:19:58 crc kubenswrapper[4948]: E0220 08:19:58.829009 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:fe85dd595906fac0fe1e7a42215bb306a963cf87d55e07cd2573726b690b2838,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kd2fq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-567668f5cf-jxwsx_openstack-operators(7c7bb531-7900-4cc6-9d9b-bae52dabc59d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 08:19:58 crc kubenswrapper[4948]: E0220 08:19:58.830231 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-jxwsx" podUID="7c7bb531-7900-4cc6-9d9b-bae52dabc59d" Feb 20 08:19:59 crc kubenswrapper[4948]: E0220 08:19:59.226545 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Feb 20 08:19:59 crc kubenswrapper[4948]: E0220 08:19:59.226766 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5r6fn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-59pcs_openstack-operators(fe0aca33-1acf-463a-91a6-bbf35a38fd7c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 08:19:59 crc kubenswrapper[4948]: E0220 08:19:59.228068 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-59pcs" podUID="fe0aca33-1acf-463a-91a6-bbf35a38fd7c" Feb 20 08:19:59 crc kubenswrapper[4948]: E0220 08:19:59.595841 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:fe85dd595906fac0fe1e7a42215bb306a963cf87d55e07cd2573726b690b2838\\\"\"" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-jxwsx" podUID="7c7bb531-7900-4cc6-9d9b-bae52dabc59d" Feb 20 08:19:59 crc kubenswrapper[4948]: E0220 08:19:59.596172 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-59pcs" podUID="fe0aca33-1acf-463a-91a6-bbf35a38fd7c" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.658508 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-6cr5p" event={"ID":"2d303d1a-1062-4bf6-be68-1bd6d1a3228f","Type":"ContainerStarted","Data":"8b305683e46bb93e0f8a4d5f84f0b8128bce46b9e3a07adf660e42f5e58cf57a"} Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.659164 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-6cr5p" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.659736 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-lj2sz" event={"ID":"f762ec0d-e09c-4baf-9540-f4ac61ce7234","Type":"ContainerStarted","Data":"05920385456f92781604cda9d67c82eb69149c95af01964924e579fb5b7a2bad"} Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.659880 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-lj2sz" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.661375 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-c6xqd" event={"ID":"586113f2-38b6-4bd1-8adb-3c155bb35ba9","Type":"ContainerStarted","Data":"1515339e6c116f2cec374cfee4da4969c8f515a00f2e4fa528d8d741cb4b08e1"} Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.661528 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-c6xqd" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.663029 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-pndq7" event={"ID":"f2400173-aa87-476c-8216-4f8c9cf9d474","Type":"ContainerStarted","Data":"01248ae14366e594770688ef93486b9c9108a6d178a5ad140b00869280852830"} Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.663232 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-pndq7" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.664617 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-kjpkr" event={"ID":"98a1aa06-948b-4034-bc07-7e546e341a8f","Type":"ContainerStarted","Data":"0deb0e674afe45a81be5d52d6248d3995b7a4749e1c63e089ef0f348365dc4d0"} Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.664749 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-kjpkr" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.667349 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-zcccd" event={"ID":"6e8e8b87-5a00-486f-a00c-1450c109f3b2","Type":"ContainerStarted","Data":"9e00325021ef73c28fdf090f3d6345fd4634b405a40c11e9d4ce76502f55fb95"} Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.667470 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-zcccd" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.668712 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-bkl49" event={"ID":"4a663826-d4b5-4ed3-8270-099b003390b4","Type":"ContainerStarted","Data":"4c8fb05f3bb4a17f2b52f611e5a7780744389a6989ee836a98b1d96068974f50"} Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.669375 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-bkl49" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.670835 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-4h9v5" event={"ID":"b8e12d0c-9564-4bf2-ac61-b22d2fbdf855","Type":"ContainerStarted","Data":"4a59274e08c5951083cdb76ae3d5000b333f30161a2bc14e5de38b9e7a8ef353"} Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.670917 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-4h9v5" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.672374 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-znrkz" event={"ID":"16933747-642c-45ff-9f98-9321c633826a","Type":"ContainerStarted","Data":"d2bcc64baae03b74eab2a38ab88906e846d18cb32b82555e43c23b89e7bb18f2"} Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.672547 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-znrkz" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.679943 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-6cr5p" podStartSLOduration=10.369867812 podStartE2EDuration="29.679928382s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:19:41.406767484 +0000 UTC m=+830.381262304" lastFinishedPulling="2026-02-20 08:20:00.716828024 +0000 UTC m=+849.691322874" observedRunningTime="2026-02-20 08:20:08.676171415 +0000 UTC m=+857.650666235" watchObservedRunningTime="2026-02-20 08:20:08.679928382 +0000 UTC m=+857.654423192" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.717919 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-4h9v5" podStartSLOduration=10.41010731 podStartE2EDuration="29.717900667s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:19:41.407851919 +0000 UTC m=+830.382346739" lastFinishedPulling="2026-02-20 08:20:00.715645236 +0000 UTC m=+849.690140096" observedRunningTime="2026-02-20 08:20:08.712712737 +0000 UTC m=+857.687207557" watchObservedRunningTime="2026-02-20 08:20:08.717900667 +0000 UTC m=+857.692395487" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.738799 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-zcccd" podStartSLOduration=10.72220704 podStartE2EDuration="29.738782328s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:19:41.700120973 +0000 UTC m=+830.674615793" lastFinishedPulling="2026-02-20 08:20:00.716696231 +0000 UTC m=+849.691191081" observedRunningTime="2026-02-20 08:20:08.735890401 +0000 UTC m=+857.710385221" watchObservedRunningTime="2026-02-20 08:20:08.738782328 +0000 UTC m=+857.713277148" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.758850 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-znrkz" podStartSLOduration=3.633832566 podStartE2EDuration="29.75883161s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:19:41.990524794 +0000 UTC m=+830.965019614" lastFinishedPulling="2026-02-20 08:20:08.115523838 +0000 UTC m=+857.090018658" observedRunningTime="2026-02-20 08:20:08.758722067 +0000 UTC m=+857.733216887" watchObservedRunningTime="2026-02-20 08:20:08.75883161 +0000 UTC m=+857.733326430" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.794647 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-kjpkr" podStartSLOduration=10.505604491 podStartE2EDuration="29.794624975s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:19:41.427789909 +0000 UTC m=+830.402284729" lastFinishedPulling="2026-02-20 08:20:00.716810393 +0000 UTC m=+849.691305213" observedRunningTime="2026-02-20 08:20:08.783685343 +0000 UTC m=+857.758180163" watchObservedRunningTime="2026-02-20 08:20:08.794624975 +0000 UTC m=+857.769119795" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.840499 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-bkl49" podStartSLOduration=10.878320488 podStartE2EDuration="29.840479741s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:19:41.754560058 +0000 UTC m=+830.729054878" lastFinishedPulling="2026-02-20 08:20:00.716719271 +0000 UTC m=+849.691214131" observedRunningTime="2026-02-20 08:20:08.815087936 +0000 UTC m=+857.789582756" watchObservedRunningTime="2026-02-20 08:20:08.840479741 +0000 UTC m=+857.814974561" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.841555 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-lj2sz" podStartSLOduration=10.916095548 podStartE2EDuration="29.841550506s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:19:41.79113891 +0000 UTC m=+830.765633730" lastFinishedPulling="2026-02-20 08:20:00.716593858 +0000 UTC m=+849.691088688" observedRunningTime="2026-02-20 08:20:08.840666055 +0000 UTC m=+857.815160875" watchObservedRunningTime="2026-02-20 08:20:08.841550506 +0000 UTC m=+857.816045326" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.868165 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-pndq7" podStartSLOduration=2.890135941 podStartE2EDuration="28.868141718s" podCreationTimestamp="2026-02-20 08:19:40 +0000 UTC" firstStartedPulling="2026-02-20 08:19:42.137636894 +0000 UTC m=+831.112131714" lastFinishedPulling="2026-02-20 08:20:08.115642641 +0000 UTC m=+857.090137491" observedRunningTime="2026-02-20 08:20:08.865083558 +0000 UTC m=+857.839578378" watchObservedRunningTime="2026-02-20 08:20:08.868141718 +0000 UTC m=+857.842636538" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.940939 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-c6xqd" podStartSLOduration=11.193230774 podStartE2EDuration="29.940920625s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:19:40.457184817 +0000 UTC m=+829.431679627" lastFinishedPulling="2026-02-20 08:19:59.204874618 +0000 UTC m=+848.179369478" observedRunningTime="2026-02-20 08:20:08.897118816 +0000 UTC m=+857.871613646" watchObservedRunningTime="2026-02-20 08:20:08.940920625 +0000 UTC m=+857.915415445" Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.679870 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-5fqzw" event={"ID":"105a118c-121d-4582-960f-1da9957980cb","Type":"ContainerStarted","Data":"a3e9593c44b650a8bda80c835d18ab928cf541897ad72ff38a186676ab68f51a"} Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.680653 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-5fqzw" Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.681623 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987464f4-xc74s" event={"ID":"347cdd7c-e5e8-49b1-a4f1-687b8a06b250","Type":"ContainerStarted","Data":"39750772c75580cafb02d68765f21c5eb0ff305d0b5a17c3eb5102bfe7d9adaa"} Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.681937 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987464f4-xc74s" Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.682965 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-dkkq9" event={"ID":"3ea675de-d1b0-4880-9652-eb066f6b0fb7","Type":"ContainerStarted","Data":"6eb878f02b7a4138e137c525a1a7da5d56bf0e67f0c5146cbf73c71230b4ec14"} Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.683148 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-dkkq9" Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.684296 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-ttkfm" event={"ID":"3fbbd11d-612d-479a-a34f-505d995a4871","Type":"ContainerStarted","Data":"915dd46835ad50adb4f7eb94e42e8320fcf29cc6712abcb40c93bf0df022e529"} Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.684613 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-ttkfm" Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.686154 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-5rrd9" event={"ID":"0f0cbb65-3c8e-41e6-8059-34e121de0821","Type":"ContainerStarted","Data":"87debc5d746918ca7f6810ce45f0ea2507d75c0c6d9365af94ce2ebc25747200"} Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.686489 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-5rrd9" Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.687445 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-7866795846-f7cxn" event={"ID":"0f98bc58-d0e3-405b-88fd-d8bd65f415a4","Type":"ContainerStarted","Data":"e8006548cce48d3fe0f5cc75d22075761038af6d79ccedbe3d91f697f3734ee2"} Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.687755 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-7866795846-f7cxn" Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.688714 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-68f46476f-5lfts" event={"ID":"d65e0993-11a0-4e81-963f-eeb9dcb92536","Type":"ContainerStarted","Data":"d630d43c3b24ea13aaf65ad43bba3892eb33c2f90cc3763b8fb4dbc270ae7add"} Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.689028 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-68f46476f-5lfts" Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.690201 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-hl5w7" event={"ID":"ee33bd02-a3ce-415d-9d2d-fefd383e9810","Type":"ContainerStarted","Data":"89e896c05dd7dac8b33eb14ad96e8b0e495eb55fb5db898aab1550650fce14f3"} Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.690492 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-hl5w7" Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.706751 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-5fqzw" podStartSLOduration=3.44130413 podStartE2EDuration="30.706736469s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:19:41.844857818 +0000 UTC m=+830.819352638" lastFinishedPulling="2026-02-20 08:20:09.110290157 +0000 UTC m=+858.084784977" observedRunningTime="2026-02-20 08:20:09.70288596 +0000 UTC m=+858.677380780" watchObservedRunningTime="2026-02-20 08:20:09.706736469 +0000 UTC m=+858.681231289" Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.741426 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-68f46476f-5lfts" podStartSLOduration=4.497568096 podStartE2EDuration="30.741410208s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:19:42.004640559 +0000 UTC m=+830.979135379" lastFinishedPulling="2026-02-20 08:20:08.248482651 +0000 UTC m=+857.222977491" observedRunningTime="2026-02-20 08:20:09.736601647 +0000 UTC m=+858.711096457" watchObservedRunningTime="2026-02-20 08:20:09.741410208 +0000 UTC m=+858.715905028" Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.758565 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-hl5w7" podStartSLOduration=3.359207277 podStartE2EDuration="30.758549572s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:19:41.813986867 +0000 UTC m=+830.788481687" lastFinishedPulling="2026-02-20 08:20:09.213329162 +0000 UTC m=+858.187823982" observedRunningTime="2026-02-20 08:20:09.758071471 +0000 UTC m=+858.732566291" watchObservedRunningTime="2026-02-20 08:20:09.758549572 +0000 UTC m=+858.733044412" Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.788526 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-7866795846-f7cxn" podStartSLOduration=3.800824093 podStartE2EDuration="29.788508903s" podCreationTimestamp="2026-02-20 08:19:40 +0000 UTC" firstStartedPulling="2026-02-20 08:19:42.128767859 +0000 UTC m=+831.103262679" lastFinishedPulling="2026-02-20 08:20:08.116452669 +0000 UTC m=+857.090947489" observedRunningTime="2026-02-20 08:20:09.782783801 +0000 UTC m=+858.757278621" watchObservedRunningTime="2026-02-20 08:20:09.788508903 +0000 UTC m=+858.763003723" Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.831568 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987464f4-xc74s" podStartSLOduration=13.03607257 podStartE2EDuration="30.831549014s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:19:41.409542988 +0000 UTC m=+830.384037808" lastFinishedPulling="2026-02-20 08:19:59.205019432 +0000 UTC m=+848.179514252" observedRunningTime="2026-02-20 08:20:09.830281615 +0000 UTC m=+858.804776425" watchObservedRunningTime="2026-02-20 08:20:09.831549014 +0000 UTC m=+858.806043834" Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.856002 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-ttkfm" podStartSLOduration=11.991700769 podStartE2EDuration="30.855984537s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:19:41.852362551 +0000 UTC m=+830.826857361" lastFinishedPulling="2026-02-20 08:20:00.716646309 +0000 UTC m=+849.691141129" observedRunningTime="2026-02-20 08:20:09.850562442 +0000 UTC m=+858.825057262" watchObservedRunningTime="2026-02-20 08:20:09.855984537 +0000 UTC m=+858.830479357" Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.894961 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-5rrd9" podStartSLOduration=4.596365751 podStartE2EDuration="30.894947345s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:19:41.995486828 +0000 UTC m=+830.969981638" lastFinishedPulling="2026-02-20 08:20:08.294068402 +0000 UTC m=+857.268563232" observedRunningTime="2026-02-20 08:20:09.892523219 +0000 UTC m=+858.867018039" watchObservedRunningTime="2026-02-20 08:20:09.894947345 +0000 UTC m=+858.869442165" Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.923101 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-dkkq9" podStartSLOduration=4.661345039 podStartE2EDuration="30.923084393s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:19:41.99858411 +0000 UTC m=+830.973078920" lastFinishedPulling="2026-02-20 08:20:08.260323444 +0000 UTC m=+857.234818274" observedRunningTime="2026-02-20 08:20:09.920024893 +0000 UTC m=+858.894519713" watchObservedRunningTime="2026-02-20 08:20:09.923084393 +0000 UTC m=+858.897579213" Feb 20 08:20:11 crc kubenswrapper[4948]: I0220 08:20:11.705103 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-jxwsx" event={"ID":"7c7bb531-7900-4cc6-9d9b-bae52dabc59d","Type":"ContainerStarted","Data":"6bf4067c3cc3f27ae6ff2e15a723b9f8e806372c7cfdf33156af0d24a9c1de0b"} Feb 20 08:20:11 crc kubenswrapper[4948]: I0220 08:20:11.705696 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-jxwsx" Feb 20 08:20:11 crc kubenswrapper[4948]: I0220 08:20:11.729620 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-jxwsx" podStartSLOduration=3.321213943 podStartE2EDuration="32.729581945s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:19:41.750419332 +0000 UTC m=+830.724914152" lastFinishedPulling="2026-02-20 08:20:11.158787334 +0000 UTC m=+860.133282154" observedRunningTime="2026-02-20 08:20:11.72241699 +0000 UTC m=+860.696911810" watchObservedRunningTime="2026-02-20 08:20:11.729581945 +0000 UTC m=+860.704076805" Feb 20 08:20:11 crc kubenswrapper[4948]: I0220 08:20:11.744010 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert\") pod \"infra-operator-controller-manager-79d975b745-jrlw5\" (UID: \"93b93501-acfe-4274-9a22-ca644b1d11d3\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" Feb 20 08:20:11 crc kubenswrapper[4948]: I0220 08:20:11.758798 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert\") pod \"infra-operator-controller-manager-79d975b745-jrlw5\" (UID: \"93b93501-acfe-4274-9a22-ca644b1d11d3\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" Feb 20 08:20:11 crc kubenswrapper[4948]: I0220 08:20:11.818122 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-nxfl4" Feb 20 08:20:11 crc kubenswrapper[4948]: I0220 08:20:11.825884 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" Feb 20 08:20:12 crc kubenswrapper[4948]: I0220 08:20:12.247815 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5"] Feb 20 08:20:12 crc kubenswrapper[4948]: W0220 08:20:12.266743 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93b93501_acfe_4274_9a22_ca644b1d11d3.slice/crio-56ecd22253c889793ab9c7e5698f95c754daf76da3e78e23c40c4217acbf5737 WatchSource:0}: Error finding container 56ecd22253c889793ab9c7e5698f95c754daf76da3e78e23c40c4217acbf5737: Status 404 returned error can't find the container with id 56ecd22253c889793ab9c7e5698f95c754daf76da3e78e23c40c4217acbf5737 Feb 20 08:20:12 crc kubenswrapper[4948]: I0220 08:20:12.454351 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m\" (UID: \"b8b07685-095c-4bbd-a30d-57ca59d7cbdc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" Feb 20 08:20:12 crc kubenswrapper[4948]: I0220 08:20:12.454497 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:20:12 crc kubenswrapper[4948]: I0220 08:20:12.454585 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:20:12 crc kubenswrapper[4948]: I0220 08:20:12.467444 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:20:12 crc kubenswrapper[4948]: I0220 08:20:12.467451 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:20:12 crc kubenswrapper[4948]: I0220 08:20:12.469061 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m\" (UID: \"b8b07685-095c-4bbd-a30d-57ca59d7cbdc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" Feb 20 08:20:12 crc kubenswrapper[4948]: I0220 08:20:12.709568 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-5sbbv" Feb 20 08:20:12 crc kubenswrapper[4948]: I0220 08:20:12.713886 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" event={"ID":"93b93501-acfe-4274-9a22-ca644b1d11d3","Type":"ContainerStarted","Data":"56ecd22253c889793ab9c7e5698f95c754daf76da3e78e23c40c4217acbf5737"} Feb 20 08:20:12 crc kubenswrapper[4948]: I0220 08:20:12.716717 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-jcnxd" event={"ID":"88142137-864d-4660-a688-a7dcc503851b","Type":"ContainerStarted","Data":"c6f25d4bd83aaf20517efd9ccdb7c5ba8f2abb24d30560a0ea4e0eba2e56aec0"} Feb 20 08:20:12 crc kubenswrapper[4948]: I0220 08:20:12.717668 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-jcnxd" Feb 20 08:20:12 crc kubenswrapper[4948]: I0220 08:20:12.717787 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:20:12 crc kubenswrapper[4948]: I0220 08:20:12.746649 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-jcnxd" podStartSLOduration=3.423646154 podStartE2EDuration="33.746618318s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:19:41.81413517 +0000 UTC m=+830.788629990" lastFinishedPulling="2026-02-20 08:20:12.137107334 +0000 UTC m=+861.111602154" observedRunningTime="2026-02-20 08:20:12.74280354 +0000 UTC m=+861.717298360" watchObservedRunningTime="2026-02-20 08:20:12.746618318 +0000 UTC m=+861.721113178" Feb 20 08:20:12 crc kubenswrapper[4948]: I0220 08:20:12.767932 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-c7nf5" Feb 20 08:20:12 crc kubenswrapper[4948]: I0220 08:20:12.776936 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" Feb 20 08:20:13 crc kubenswrapper[4948]: I0220 08:20:13.000309 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk"] Feb 20 08:20:13 crc kubenswrapper[4948]: W0220 08:20:13.013687 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode9f13bcd_ac5a_4cbd_952a_a5b9dceee562.slice/crio-80bd3ea908fb4eb09f8f4416da6a368cb8331471b5e64de7116e12d5a966ea56 WatchSource:0}: Error finding container 80bd3ea908fb4eb09f8f4416da6a368cb8331471b5e64de7116e12d5a966ea56: Status 404 returned error can't find the container with id 80bd3ea908fb4eb09f8f4416da6a368cb8331471b5e64de7116e12d5a966ea56 Feb 20 08:20:13 crc kubenswrapper[4948]: I0220 08:20:13.284209 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m"] Feb 20 08:20:13 crc kubenswrapper[4948]: W0220 08:20:13.300266 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb8b07685_095c_4bbd_a30d_57ca59d7cbdc.slice/crio-95d33a746d766d859f961c9d4ead205a2fa3e2416d7051db3d97c9a00f159cfa WatchSource:0}: Error finding container 95d33a746d766d859f961c9d4ead205a2fa3e2416d7051db3d97c9a00f159cfa: Status 404 returned error can't find the container with id 95d33a746d766d859f961c9d4ead205a2fa3e2416d7051db3d97c9a00f159cfa Feb 20 08:20:13 crc kubenswrapper[4948]: I0220 08:20:13.739346 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" event={"ID":"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562","Type":"ContainerStarted","Data":"0a26ac33ec17bf06b70ad74d59f7ee41394a2704c8d4b99685e4a4ce9ccf41fa"} Feb 20 08:20:13 crc kubenswrapper[4948]: I0220 08:20:13.739384 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" event={"ID":"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562","Type":"ContainerStarted","Data":"80bd3ea908fb4eb09f8f4416da6a368cb8331471b5e64de7116e12d5a966ea56"} Feb 20 08:20:13 crc kubenswrapper[4948]: I0220 08:20:13.739397 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" event={"ID":"b8b07685-095c-4bbd-a30d-57ca59d7cbdc","Type":"ContainerStarted","Data":"95d33a746d766d859f961c9d4ead205a2fa3e2416d7051db3d97c9a00f159cfa"} Feb 20 08:20:13 crc kubenswrapper[4948]: I0220 08:20:13.739410 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-59pcs" event={"ID":"fe0aca33-1acf-463a-91a6-bbf35a38fd7c","Type":"ContainerStarted","Data":"b5e51c8afb7e870e869321f303b239d0d185612855b1179878e8aaf17edaffcb"} Feb 20 08:20:13 crc kubenswrapper[4948]: I0220 08:20:13.739426 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:20:13 crc kubenswrapper[4948]: I0220 08:20:13.755931 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" podStartSLOduration=33.755916171 podStartE2EDuration="33.755916171s" podCreationTimestamp="2026-02-20 08:19:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:20:13.75459106 +0000 UTC m=+862.729085910" watchObservedRunningTime="2026-02-20 08:20:13.755916171 +0000 UTC m=+862.730410991" Feb 20 08:20:13 crc kubenswrapper[4948]: I0220 08:20:13.784409 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-59pcs" podStartSLOduration=2.582149195 podStartE2EDuration="33.784391917s" podCreationTimestamp="2026-02-20 08:19:40 +0000 UTC" firstStartedPulling="2026-02-20 08:19:42.106879305 +0000 UTC m=+831.081374125" lastFinishedPulling="2026-02-20 08:20:13.309122007 +0000 UTC m=+862.283616847" observedRunningTime="2026-02-20 08:20:13.777319044 +0000 UTC m=+862.751813874" watchObservedRunningTime="2026-02-20 08:20:13.784391917 +0000 UTC m=+862.758886747" Feb 20 08:20:14 crc kubenswrapper[4948]: I0220 08:20:14.744114 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" event={"ID":"93b93501-acfe-4274-9a22-ca644b1d11d3","Type":"ContainerStarted","Data":"118b2eef7d5016594db3982e89ab763d690309404c4d622207ee663120afdae1"} Feb 20 08:20:14 crc kubenswrapper[4948]: I0220 08:20:14.768250 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" podStartSLOduration=33.66335475 podStartE2EDuration="35.768229535s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:20:12.26928645 +0000 UTC m=+861.243781270" lastFinishedPulling="2026-02-20 08:20:14.374161235 +0000 UTC m=+863.348656055" observedRunningTime="2026-02-20 08:20:14.757200981 +0000 UTC m=+863.731695811" watchObservedRunningTime="2026-02-20 08:20:14.768229535 +0000 UTC m=+863.742724355" Feb 20 08:20:15 crc kubenswrapper[4948]: I0220 08:20:15.751849 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" Feb 20 08:20:16 crc kubenswrapper[4948]: I0220 08:20:16.761573 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" event={"ID":"b8b07685-095c-4bbd-a30d-57ca59d7cbdc","Type":"ContainerStarted","Data":"3e13e2e5df6fe5e1ffd0e5435f05a87967318a31abfc6a13ad7a56936b7cec37"} Feb 20 08:20:16 crc kubenswrapper[4948]: I0220 08:20:16.762061 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" Feb 20 08:20:16 crc kubenswrapper[4948]: I0220 08:20:16.804920 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" podStartSLOduration=35.346814965 podStartE2EDuration="37.804897129s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:20:13.309294291 +0000 UTC m=+862.283789111" lastFinishedPulling="2026-02-20 08:20:15.767376445 +0000 UTC m=+864.741871275" observedRunningTime="2026-02-20 08:20:16.796966206 +0000 UTC m=+865.771461076" watchObservedRunningTime="2026-02-20 08:20:16.804897129 +0000 UTC m=+865.779391949" Feb 20 08:20:19 crc kubenswrapper[4948]: I0220 08:20:19.873269 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-4h9v5" Feb 20 08:20:19 crc kubenswrapper[4948]: I0220 08:20:19.886738 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-c6xqd" Feb 20 08:20:19 crc kubenswrapper[4948]: I0220 08:20:19.907781 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-kjpkr" Feb 20 08:20:19 crc kubenswrapper[4948]: I0220 08:20:19.974460 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-6cr5p" Feb 20 08:20:19 crc kubenswrapper[4948]: I0220 08:20:19.986887 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987464f4-xc74s" Feb 20 08:20:20 crc kubenswrapper[4948]: I0220 08:20:20.035258 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-bkl49" Feb 20 08:20:20 crc kubenswrapper[4948]: I0220 08:20:20.058379 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-5fqzw" Feb 20 08:20:20 crc kubenswrapper[4948]: I0220 08:20:20.090740 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-lj2sz" Feb 20 08:20:20 crc kubenswrapper[4948]: I0220 08:20:20.158634 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-ttkfm" Feb 20 08:20:20 crc kubenswrapper[4948]: I0220 08:20:20.181225 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-zcccd" Feb 20 08:20:20 crc kubenswrapper[4948]: I0220 08:20:20.263132 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-jcnxd" Feb 20 08:20:20 crc kubenswrapper[4948]: I0220 08:20:20.286947 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-jxwsx" Feb 20 08:20:20 crc kubenswrapper[4948]: I0220 08:20:20.323616 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-hl5w7" Feb 20 08:20:20 crc kubenswrapper[4948]: I0220 08:20:20.400287 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-dkkq9" Feb 20 08:20:20 crc kubenswrapper[4948]: I0220 08:20:20.422822 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-znrkz" Feb 20 08:20:20 crc kubenswrapper[4948]: I0220 08:20:20.454124 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-68f46476f-5lfts" Feb 20 08:20:20 crc kubenswrapper[4948]: I0220 08:20:20.503887 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-5rrd9" Feb 20 08:20:20 crc kubenswrapper[4948]: I0220 08:20:20.558776 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-7866795846-f7cxn" Feb 20 08:20:20 crc kubenswrapper[4948]: I0220 08:20:20.572848 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-pndq7" Feb 20 08:20:21 crc kubenswrapper[4948]: I0220 08:20:21.836534 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" Feb 20 08:20:22 crc kubenswrapper[4948]: I0220 08:20:22.726030 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:20:22 crc kubenswrapper[4948]: I0220 08:20:22.794334 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" Feb 20 08:20:38 crc kubenswrapper[4948]: I0220 08:20:38.025017 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:20:38 crc kubenswrapper[4948]: I0220 08:20:38.025472 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.757659 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9dfhs"] Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.759022 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-9dfhs" Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.763467 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.763682 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.763827 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.764033 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-qgm2g" Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.777877 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9dfhs"] Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.815524 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-cqgz2"] Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.816561 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-cqgz2" Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.819804 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.835203 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-cqgz2"] Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.890404 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7d754\" (UniqueName: \"kubernetes.io/projected/1224cf30-813c-48f2-92c9-fabe78bbc6bc-kube-api-access-7d754\") pod \"dnsmasq-dns-675f4bcbfc-9dfhs\" (UID: \"1224cf30-813c-48f2-92c9-fabe78bbc6bc\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9dfhs" Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.890599 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1224cf30-813c-48f2-92c9-fabe78bbc6bc-config\") pod \"dnsmasq-dns-675f4bcbfc-9dfhs\" (UID: \"1224cf30-813c-48f2-92c9-fabe78bbc6bc\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9dfhs" Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.991552 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7d754\" (UniqueName: \"kubernetes.io/projected/1224cf30-813c-48f2-92c9-fabe78bbc6bc-kube-api-access-7d754\") pod \"dnsmasq-dns-675f4bcbfc-9dfhs\" (UID: \"1224cf30-813c-48f2-92c9-fabe78bbc6bc\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9dfhs" Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.991632 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffpsp\" (UniqueName: \"kubernetes.io/projected/05b1d586-95db-4795-a032-f22ba90e6c16-kube-api-access-ffpsp\") pod \"dnsmasq-dns-78dd6ddcc-cqgz2\" (UID: \"05b1d586-95db-4795-a032-f22ba90e6c16\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cqgz2" Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.991680 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05b1d586-95db-4795-a032-f22ba90e6c16-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-cqgz2\" (UID: \"05b1d586-95db-4795-a032-f22ba90e6c16\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cqgz2" Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.991703 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05b1d586-95db-4795-a032-f22ba90e6c16-config\") pod \"dnsmasq-dns-78dd6ddcc-cqgz2\" (UID: \"05b1d586-95db-4795-a032-f22ba90e6c16\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cqgz2" Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.991743 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1224cf30-813c-48f2-92c9-fabe78bbc6bc-config\") pod \"dnsmasq-dns-675f4bcbfc-9dfhs\" (UID: \"1224cf30-813c-48f2-92c9-fabe78bbc6bc\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9dfhs" Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.992517 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1224cf30-813c-48f2-92c9-fabe78bbc6bc-config\") pod \"dnsmasq-dns-675f4bcbfc-9dfhs\" (UID: \"1224cf30-813c-48f2-92c9-fabe78bbc6bc\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9dfhs" Feb 20 08:20:40 crc kubenswrapper[4948]: I0220 08:20:40.009194 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7d754\" (UniqueName: \"kubernetes.io/projected/1224cf30-813c-48f2-92c9-fabe78bbc6bc-kube-api-access-7d754\") pod \"dnsmasq-dns-675f4bcbfc-9dfhs\" (UID: \"1224cf30-813c-48f2-92c9-fabe78bbc6bc\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9dfhs" Feb 20 08:20:40 crc kubenswrapper[4948]: I0220 08:20:40.080862 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-9dfhs" Feb 20 08:20:40 crc kubenswrapper[4948]: I0220 08:20:40.093360 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffpsp\" (UniqueName: \"kubernetes.io/projected/05b1d586-95db-4795-a032-f22ba90e6c16-kube-api-access-ffpsp\") pod \"dnsmasq-dns-78dd6ddcc-cqgz2\" (UID: \"05b1d586-95db-4795-a032-f22ba90e6c16\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cqgz2" Feb 20 08:20:40 crc kubenswrapper[4948]: I0220 08:20:40.093451 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05b1d586-95db-4795-a032-f22ba90e6c16-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-cqgz2\" (UID: \"05b1d586-95db-4795-a032-f22ba90e6c16\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cqgz2" Feb 20 08:20:40 crc kubenswrapper[4948]: I0220 08:20:40.093489 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05b1d586-95db-4795-a032-f22ba90e6c16-config\") pod \"dnsmasq-dns-78dd6ddcc-cqgz2\" (UID: \"05b1d586-95db-4795-a032-f22ba90e6c16\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cqgz2" Feb 20 08:20:40 crc kubenswrapper[4948]: I0220 08:20:40.094566 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05b1d586-95db-4795-a032-f22ba90e6c16-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-cqgz2\" (UID: \"05b1d586-95db-4795-a032-f22ba90e6c16\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cqgz2" Feb 20 08:20:40 crc kubenswrapper[4948]: I0220 08:20:40.094565 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05b1d586-95db-4795-a032-f22ba90e6c16-config\") pod \"dnsmasq-dns-78dd6ddcc-cqgz2\" (UID: \"05b1d586-95db-4795-a032-f22ba90e6c16\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cqgz2" Feb 20 08:20:40 crc kubenswrapper[4948]: I0220 08:20:40.125916 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffpsp\" (UniqueName: \"kubernetes.io/projected/05b1d586-95db-4795-a032-f22ba90e6c16-kube-api-access-ffpsp\") pod \"dnsmasq-dns-78dd6ddcc-cqgz2\" (UID: \"05b1d586-95db-4795-a032-f22ba90e6c16\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cqgz2" Feb 20 08:20:40 crc kubenswrapper[4948]: I0220 08:20:40.133963 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-cqgz2" Feb 20 08:20:40 crc kubenswrapper[4948]: I0220 08:20:40.598701 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9dfhs"] Feb 20 08:20:40 crc kubenswrapper[4948]: I0220 08:20:40.599947 4948 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 08:20:40 crc kubenswrapper[4948]: W0220 08:20:40.608279 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod05b1d586_95db_4795_a032_f22ba90e6c16.slice/crio-4e5e167063afcb1f5336b93cee715fd831f959a4a5fb12e9e562274ec18ad672 WatchSource:0}: Error finding container 4e5e167063afcb1f5336b93cee715fd831f959a4a5fb12e9e562274ec18ad672: Status 404 returned error can't find the container with id 4e5e167063afcb1f5336b93cee715fd831f959a4a5fb12e9e562274ec18ad672 Feb 20 08:20:40 crc kubenswrapper[4948]: I0220 08:20:40.612651 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-cqgz2"] Feb 20 08:20:40 crc kubenswrapper[4948]: I0220 08:20:40.969921 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-9dfhs" event={"ID":"1224cf30-813c-48f2-92c9-fabe78bbc6bc","Type":"ContainerStarted","Data":"459736be569ebf9acfc7e517d56782883ab07764b913b8c4d305735259e95b55"} Feb 20 08:20:40 crc kubenswrapper[4948]: I0220 08:20:40.971360 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-cqgz2" event={"ID":"05b1d586-95db-4795-a032-f22ba90e6c16","Type":"ContainerStarted","Data":"4e5e167063afcb1f5336b93cee715fd831f959a4a5fb12e9e562274ec18ad672"} Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.526616 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9dfhs"] Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.564921 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zt6gg"] Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.566566 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.572546 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zt6gg"] Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.642457 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pds7s\" (UniqueName: \"kubernetes.io/projected/977992e9-30ca-4420-98ae-92f91545cd10-kube-api-access-pds7s\") pod \"dnsmasq-dns-666b6646f7-zt6gg\" (UID: \"977992e9-30ca-4420-98ae-92f91545cd10\") " pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.642941 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/977992e9-30ca-4420-98ae-92f91545cd10-dns-svc\") pod \"dnsmasq-dns-666b6646f7-zt6gg\" (UID: \"977992e9-30ca-4420-98ae-92f91545cd10\") " pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.643076 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/977992e9-30ca-4420-98ae-92f91545cd10-config\") pod \"dnsmasq-dns-666b6646f7-zt6gg\" (UID: \"977992e9-30ca-4420-98ae-92f91545cd10\") " pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.743863 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pds7s\" (UniqueName: \"kubernetes.io/projected/977992e9-30ca-4420-98ae-92f91545cd10-kube-api-access-pds7s\") pod \"dnsmasq-dns-666b6646f7-zt6gg\" (UID: \"977992e9-30ca-4420-98ae-92f91545cd10\") " pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.743925 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/977992e9-30ca-4420-98ae-92f91545cd10-dns-svc\") pod \"dnsmasq-dns-666b6646f7-zt6gg\" (UID: \"977992e9-30ca-4420-98ae-92f91545cd10\") " pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.743989 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/977992e9-30ca-4420-98ae-92f91545cd10-config\") pod \"dnsmasq-dns-666b6646f7-zt6gg\" (UID: \"977992e9-30ca-4420-98ae-92f91545cd10\") " pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.745146 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/977992e9-30ca-4420-98ae-92f91545cd10-config\") pod \"dnsmasq-dns-666b6646f7-zt6gg\" (UID: \"977992e9-30ca-4420-98ae-92f91545cd10\") " pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.746071 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/977992e9-30ca-4420-98ae-92f91545cd10-dns-svc\") pod \"dnsmasq-dns-666b6646f7-zt6gg\" (UID: \"977992e9-30ca-4420-98ae-92f91545cd10\") " pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.765593 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pds7s\" (UniqueName: \"kubernetes.io/projected/977992e9-30ca-4420-98ae-92f91545cd10-kube-api-access-pds7s\") pod \"dnsmasq-dns-666b6646f7-zt6gg\" (UID: \"977992e9-30ca-4420-98ae-92f91545cd10\") " pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.879069 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-cqgz2"] Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.895850 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.921143 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-mkkzf"] Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.927271 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.941611 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-mkkzf"] Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.048022 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdnm9\" (UniqueName: \"kubernetes.io/projected/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-kube-api-access-hdnm9\") pod \"dnsmasq-dns-57d769cc4f-mkkzf\" (UID: \"5686ae39-4e44-4303-9b44-ad4b9ceae1ba\") " pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.048096 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-config\") pod \"dnsmasq-dns-57d769cc4f-mkkzf\" (UID: \"5686ae39-4e44-4303-9b44-ad4b9ceae1ba\") " pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.048146 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-mkkzf\" (UID: \"5686ae39-4e44-4303-9b44-ad4b9ceae1ba\") " pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.149790 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-config\") pod \"dnsmasq-dns-57d769cc4f-mkkzf\" (UID: \"5686ae39-4e44-4303-9b44-ad4b9ceae1ba\") " pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.149864 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-mkkzf\" (UID: \"5686ae39-4e44-4303-9b44-ad4b9ceae1ba\") " pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.149914 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdnm9\" (UniqueName: \"kubernetes.io/projected/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-kube-api-access-hdnm9\") pod \"dnsmasq-dns-57d769cc4f-mkkzf\" (UID: \"5686ae39-4e44-4303-9b44-ad4b9ceae1ba\") " pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.150868 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-config\") pod \"dnsmasq-dns-57d769cc4f-mkkzf\" (UID: \"5686ae39-4e44-4303-9b44-ad4b9ceae1ba\") " pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.151096 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-mkkzf\" (UID: \"5686ae39-4e44-4303-9b44-ad4b9ceae1ba\") " pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.173997 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdnm9\" (UniqueName: \"kubernetes.io/projected/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-kube-api-access-hdnm9\") pod \"dnsmasq-dns-57d769cc4f-mkkzf\" (UID: \"5686ae39-4e44-4303-9b44-ad4b9ceae1ba\") " pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.316471 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.424430 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zt6gg"] Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.774450 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.776295 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.778663 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.783628 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.783833 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.783951 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.784072 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-8zcmc" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.784175 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.784234 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.797520 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.867695 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-mkkzf"] Feb 20 08:20:43 crc kubenswrapper[4948]: W0220 08:20:43.882387 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5686ae39_4e44_4303_9b44_ad4b9ceae1ba.slice/crio-b4980c083e552022d3673e8c44cd0e8286ee33627cbef5030e75bd7be40133d3 WatchSource:0}: Error finding container b4980c083e552022d3673e8c44cd0e8286ee33627cbef5030e75bd7be40133d3: Status 404 returned error can't find the container with id b4980c083e552022d3673e8c44cd0e8286ee33627cbef5030e75bd7be40133d3 Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.963412 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/73b84bb7-f594-4823-ac03-40fdac6ee177-pod-info\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.963520 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqlcq\" (UniqueName: \"kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-kube-api-access-tqlcq\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.963663 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.963709 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.963825 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/73b84bb7-f594-4823-ac03-40fdac6ee177-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.964835 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-server-conf\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.964918 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.965133 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-config-data\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.965228 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.965330 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.965347 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.006642 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" event={"ID":"5686ae39-4e44-4303-9b44-ad4b9ceae1ba","Type":"ContainerStarted","Data":"b4980c083e552022d3673e8c44cd0e8286ee33627cbef5030e75bd7be40133d3"} Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.008015 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" event={"ID":"977992e9-30ca-4420-98ae-92f91545cd10","Type":"ContainerStarted","Data":"a0508d6a1f4b7f39846aac7f90bdfe83ad8793d4225f8bb5f2eaca0710969807"} Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.068987 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.069047 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.069063 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.069092 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/73b84bb7-f594-4823-ac03-40fdac6ee177-pod-info\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.069123 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqlcq\" (UniqueName: \"kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-kube-api-access-tqlcq\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.069139 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.069162 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.069178 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/73b84bb7-f594-4823-ac03-40fdac6ee177-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.069200 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-server-conf\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.069226 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.069251 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-config-data\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.069477 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.069910 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.070699 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.070748 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-config-data\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.070880 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.071231 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-server-conf\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.079019 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/73b84bb7-f594-4823-ac03-40fdac6ee177-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.080674 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/73b84bb7-f594-4823-ac03-40fdac6ee177-pod-info\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.082032 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.084316 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.100691 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqlcq\" (UniqueName: \"kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-kube-api-access-tqlcq\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.107661 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.116465 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.132486 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.145886 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.146007 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.149255 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.149449 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.149560 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-rrjt5" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.149663 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.149758 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.149873 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.150067 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.203624 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.203708 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.203749 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.203790 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.203812 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.203845 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.203879 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7t2f\" (UniqueName: \"kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-kube-api-access-b7t2f\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.203904 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.203946 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/790894c9-053c-497c-955d-ce7519111dd6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.203964 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.204889 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/790894c9-053c-497c-955d-ce7519111dd6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.309839 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.309901 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.309941 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.309998 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7t2f\" (UniqueName: \"kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-kube-api-access-b7t2f\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.310030 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.310058 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/790894c9-053c-497c-955d-ce7519111dd6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.310087 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.310111 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/790894c9-053c-497c-955d-ce7519111dd6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.310147 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.310207 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.310247 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.310638 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.311291 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.311865 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.312108 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.313112 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.313149 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.313399 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.317290 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/790894c9-053c-497c-955d-ce7519111dd6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.317461 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/790894c9-053c-497c-955d-ce7519111dd6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.330722 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.331723 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.331845 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7t2f\" (UniqueName: \"kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-kube-api-access-b7t2f\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.397123 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 20 08:20:44 crc kubenswrapper[4948]: W0220 08:20:44.403661 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73b84bb7_f594_4823_ac03_40fdac6ee177.slice/crio-e552b8693d7f82c1ed5a2c7c103ec119814c444b37d5f93e2491692378454a4c WatchSource:0}: Error finding container e552b8693d7f82c1ed5a2c7c103ec119814c444b37d5f93e2491692378454a4c: Status 404 returned error can't find the container with id e552b8693d7f82c1ed5a2c7c103ec119814c444b37d5f93e2491692378454a4c Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.496737 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.027931 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"73b84bb7-f594-4823-ac03-40fdac6ee177","Type":"ContainerStarted","Data":"e552b8693d7f82c1ed5a2c7c103ec119814c444b37d5f93e2491692378454a4c"} Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.186081 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.191312 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.194287 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.194407 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-w44rf" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.194534 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.194868 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.200065 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.202025 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.325407 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6e59b63d-0c56-488c-87cb-348af87058c6-kolla-config\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.325503 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6e59b63d-0c56-488c-87cb-348af87058c6-config-data-generated\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.325524 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e59b63d-0c56-488c-87cb-348af87058c6-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.325564 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e59b63d-0c56-488c-87cb-348af87058c6-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.325593 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zq68c\" (UniqueName: \"kubernetes.io/projected/6e59b63d-0c56-488c-87cb-348af87058c6-kube-api-access-zq68c\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.325610 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.325635 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6e59b63d-0c56-488c-87cb-348af87058c6-config-data-default\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.325654 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e59b63d-0c56-488c-87cb-348af87058c6-operator-scripts\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.431688 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6e59b63d-0c56-488c-87cb-348af87058c6-kolla-config\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.431963 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6e59b63d-0c56-488c-87cb-348af87058c6-config-data-generated\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.431996 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e59b63d-0c56-488c-87cb-348af87058c6-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.432042 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e59b63d-0c56-488c-87cb-348af87058c6-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.432077 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zq68c\" (UniqueName: \"kubernetes.io/projected/6e59b63d-0c56-488c-87cb-348af87058c6-kube-api-access-zq68c\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.432114 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.432147 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6e59b63d-0c56-488c-87cb-348af87058c6-config-data-default\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.432165 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e59b63d-0c56-488c-87cb-348af87058c6-operator-scripts\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.432348 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rt8tp"] Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.434047 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rt8tp" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.436565 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.437301 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e59b63d-0c56-488c-87cb-348af87058c6-operator-scripts\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.437537 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6e59b63d-0c56-488c-87cb-348af87058c6-config-data-default\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.437542 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6e59b63d-0c56-488c-87cb-348af87058c6-kolla-config\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.440215 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6e59b63d-0c56-488c-87cb-348af87058c6-config-data-generated\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.464657 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rt8tp"] Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.481852 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e59b63d-0c56-488c-87cb-348af87058c6-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.483259 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zq68c\" (UniqueName: \"kubernetes.io/projected/6e59b63d-0c56-488c-87cb-348af87058c6-kube-api-access-zq68c\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.492028 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.498276 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e59b63d-0c56-488c-87cb-348af87058c6-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.518315 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.635050 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93f24735-d27d-4867-b9f7-cce47e7f376f-utilities\") pod \"redhat-operators-rt8tp\" (UID: \"93f24735-d27d-4867-b9f7-cce47e7f376f\") " pod="openshift-marketplace/redhat-operators-rt8tp" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.636641 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2cfc\" (UniqueName: \"kubernetes.io/projected/93f24735-d27d-4867-b9f7-cce47e7f376f-kube-api-access-q2cfc\") pod \"redhat-operators-rt8tp\" (UID: \"93f24735-d27d-4867-b9f7-cce47e7f376f\") " pod="openshift-marketplace/redhat-operators-rt8tp" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.636704 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93f24735-d27d-4867-b9f7-cce47e7f376f-catalog-content\") pod \"redhat-operators-rt8tp\" (UID: \"93f24735-d27d-4867-b9f7-cce47e7f376f\") " pod="openshift-marketplace/redhat-operators-rt8tp" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.738346 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93f24735-d27d-4867-b9f7-cce47e7f376f-utilities\") pod \"redhat-operators-rt8tp\" (UID: \"93f24735-d27d-4867-b9f7-cce47e7f376f\") " pod="openshift-marketplace/redhat-operators-rt8tp" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.738400 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2cfc\" (UniqueName: \"kubernetes.io/projected/93f24735-d27d-4867-b9f7-cce47e7f376f-kube-api-access-q2cfc\") pod \"redhat-operators-rt8tp\" (UID: \"93f24735-d27d-4867-b9f7-cce47e7f376f\") " pod="openshift-marketplace/redhat-operators-rt8tp" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.738444 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93f24735-d27d-4867-b9f7-cce47e7f376f-catalog-content\") pod \"redhat-operators-rt8tp\" (UID: \"93f24735-d27d-4867-b9f7-cce47e7f376f\") " pod="openshift-marketplace/redhat-operators-rt8tp" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.739092 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93f24735-d27d-4867-b9f7-cce47e7f376f-utilities\") pod \"redhat-operators-rt8tp\" (UID: \"93f24735-d27d-4867-b9f7-cce47e7f376f\") " pod="openshift-marketplace/redhat-operators-rt8tp" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.739126 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93f24735-d27d-4867-b9f7-cce47e7f376f-catalog-content\") pod \"redhat-operators-rt8tp\" (UID: \"93f24735-d27d-4867-b9f7-cce47e7f376f\") " pod="openshift-marketplace/redhat-operators-rt8tp" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.757178 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2cfc\" (UniqueName: \"kubernetes.io/projected/93f24735-d27d-4867-b9f7-cce47e7f376f-kube-api-access-q2cfc\") pod \"redhat-operators-rt8tp\" (UID: \"93f24735-d27d-4867-b9f7-cce47e7f376f\") " pod="openshift-marketplace/redhat-operators-rt8tp" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.852465 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rt8tp" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.562889 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.564163 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.574291 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-wmf44" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.574688 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.575040 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.575178 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.582064 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.654992 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qwvn\" (UniqueName: \"kubernetes.io/projected/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-kube-api-access-7qwvn\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.655069 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.655102 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.655124 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.655139 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.655159 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.655237 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.655273 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.756697 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.756990 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.757031 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qwvn\" (UniqueName: \"kubernetes.io/projected/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-kube-api-access-7qwvn\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.757067 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.757097 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.757119 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.757135 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.757157 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.757834 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.758071 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.760199 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.760541 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.763526 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.763769 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.768983 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.769865 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.778018 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.780220 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.780401 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.780506 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-4mvzk" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.781431 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.795103 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.805439 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qwvn\" (UniqueName: \"kubernetes.io/projected/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-kube-api-access-7qwvn\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.858522 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8e994489-a2cf-4f6d-a00c-98f627ba0e5f-config-data\") pod \"memcached-0\" (UID: \"8e994489-a2cf-4f6d-a00c-98f627ba0e5f\") " pod="openstack/memcached-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.858602 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mngjx\" (UniqueName: \"kubernetes.io/projected/8e994489-a2cf-4f6d-a00c-98f627ba0e5f-kube-api-access-mngjx\") pod \"memcached-0\" (UID: \"8e994489-a2cf-4f6d-a00c-98f627ba0e5f\") " pod="openstack/memcached-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.858629 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e994489-a2cf-4f6d-a00c-98f627ba0e5f-memcached-tls-certs\") pod \"memcached-0\" (UID: \"8e994489-a2cf-4f6d-a00c-98f627ba0e5f\") " pod="openstack/memcached-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.858647 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e994489-a2cf-4f6d-a00c-98f627ba0e5f-combined-ca-bundle\") pod \"memcached-0\" (UID: \"8e994489-a2cf-4f6d-a00c-98f627ba0e5f\") " pod="openstack/memcached-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.858673 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8e994489-a2cf-4f6d-a00c-98f627ba0e5f-kolla-config\") pod \"memcached-0\" (UID: \"8e994489-a2cf-4f6d-a00c-98f627ba0e5f\") " pod="openstack/memcached-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.897168 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.959479 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8e994489-a2cf-4f6d-a00c-98f627ba0e5f-config-data\") pod \"memcached-0\" (UID: \"8e994489-a2cf-4f6d-a00c-98f627ba0e5f\") " pod="openstack/memcached-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.959544 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mngjx\" (UniqueName: \"kubernetes.io/projected/8e994489-a2cf-4f6d-a00c-98f627ba0e5f-kube-api-access-mngjx\") pod \"memcached-0\" (UID: \"8e994489-a2cf-4f6d-a00c-98f627ba0e5f\") " pod="openstack/memcached-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.959574 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e994489-a2cf-4f6d-a00c-98f627ba0e5f-memcached-tls-certs\") pod \"memcached-0\" (UID: \"8e994489-a2cf-4f6d-a00c-98f627ba0e5f\") " pod="openstack/memcached-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.959592 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e994489-a2cf-4f6d-a00c-98f627ba0e5f-combined-ca-bundle\") pod \"memcached-0\" (UID: \"8e994489-a2cf-4f6d-a00c-98f627ba0e5f\") " pod="openstack/memcached-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.959610 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8e994489-a2cf-4f6d-a00c-98f627ba0e5f-kolla-config\") pod \"memcached-0\" (UID: \"8e994489-a2cf-4f6d-a00c-98f627ba0e5f\") " pod="openstack/memcached-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.960248 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8e994489-a2cf-4f6d-a00c-98f627ba0e5f-kolla-config\") pod \"memcached-0\" (UID: \"8e994489-a2cf-4f6d-a00c-98f627ba0e5f\") " pod="openstack/memcached-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.960586 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8e994489-a2cf-4f6d-a00c-98f627ba0e5f-config-data\") pod \"memcached-0\" (UID: \"8e994489-a2cf-4f6d-a00c-98f627ba0e5f\") " pod="openstack/memcached-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.962991 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e994489-a2cf-4f6d-a00c-98f627ba0e5f-memcached-tls-certs\") pod \"memcached-0\" (UID: \"8e994489-a2cf-4f6d-a00c-98f627ba0e5f\") " pod="openstack/memcached-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.966596 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e994489-a2cf-4f6d-a00c-98f627ba0e5f-combined-ca-bundle\") pod \"memcached-0\" (UID: \"8e994489-a2cf-4f6d-a00c-98f627ba0e5f\") " pod="openstack/memcached-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.976118 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mngjx\" (UniqueName: \"kubernetes.io/projected/8e994489-a2cf-4f6d-a00c-98f627ba0e5f-kube-api-access-mngjx\") pod \"memcached-0\" (UID: \"8e994489-a2cf-4f6d-a00c-98f627ba0e5f\") " pod="openstack/memcached-0" Feb 20 08:20:47 crc kubenswrapper[4948]: I0220 08:20:47.147203 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 20 08:20:49 crc kubenswrapper[4948]: I0220 08:20:49.040986 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 20 08:20:49 crc kubenswrapper[4948]: I0220 08:20:49.042380 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 20 08:20:49 crc kubenswrapper[4948]: I0220 08:20:49.044748 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-d57tg" Feb 20 08:20:49 crc kubenswrapper[4948]: I0220 08:20:49.052341 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 20 08:20:49 crc kubenswrapper[4948]: I0220 08:20:49.204133 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqgmb\" (UniqueName: \"kubernetes.io/projected/0dfbb200-e444-4895-a410-e51cbb1fc112-kube-api-access-rqgmb\") pod \"kube-state-metrics-0\" (UID: \"0dfbb200-e444-4895-a410-e51cbb1fc112\") " pod="openstack/kube-state-metrics-0" Feb 20 08:20:49 crc kubenswrapper[4948]: I0220 08:20:49.305693 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqgmb\" (UniqueName: \"kubernetes.io/projected/0dfbb200-e444-4895-a410-e51cbb1fc112-kube-api-access-rqgmb\") pod \"kube-state-metrics-0\" (UID: \"0dfbb200-e444-4895-a410-e51cbb1fc112\") " pod="openstack/kube-state-metrics-0" Feb 20 08:20:49 crc kubenswrapper[4948]: I0220 08:20:49.326673 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqgmb\" (UniqueName: \"kubernetes.io/projected/0dfbb200-e444-4895-a410-e51cbb1fc112-kube-api-access-rqgmb\") pod \"kube-state-metrics-0\" (UID: \"0dfbb200-e444-4895-a410-e51cbb1fc112\") " pod="openstack/kube-state-metrics-0" Feb 20 08:20:49 crc kubenswrapper[4948]: I0220 08:20:49.373628 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.844697 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-mkpng"] Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.845756 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-mkpng" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.851278 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-258bh" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.851595 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.854076 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.854081 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-q4bv4"] Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.856694 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.873295 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-q4bv4"] Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.890033 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-mkpng"] Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.953901 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/c82772b6-9737-45ee-998a-e897086c03b5-var-log\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.953965 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/b80cb988-de2a-4e65-9161-a0af0561c754-ovn-controller-tls-certs\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.954012 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b80cb988-de2a-4e65-9161-a0af0561c754-var-run\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.954088 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b80cb988-de2a-4e65-9161-a0af0561c754-combined-ca-bundle\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.954251 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c82772b6-9737-45ee-998a-e897086c03b5-var-run\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.954326 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b80cb988-de2a-4e65-9161-a0af0561c754-var-log-ovn\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.954356 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b80cb988-de2a-4e65-9161-a0af0561c754-var-run-ovn\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.954383 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c82772b6-9737-45ee-998a-e897086c03b5-scripts\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.954424 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b80cb988-de2a-4e65-9161-a0af0561c754-scripts\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.954461 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hn4rx\" (UniqueName: \"kubernetes.io/projected/c82772b6-9737-45ee-998a-e897086c03b5-kube-api-access-hn4rx\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.954524 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/c82772b6-9737-45ee-998a-e897086c03b5-etc-ovs\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.954657 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5ql2\" (UniqueName: \"kubernetes.io/projected/b80cb988-de2a-4e65-9161-a0af0561c754-kube-api-access-h5ql2\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.954732 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/c82772b6-9737-45ee-998a-e897086c03b5-var-lib\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.056417 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hn4rx\" (UniqueName: \"kubernetes.io/projected/c82772b6-9737-45ee-998a-e897086c03b5-kube-api-access-hn4rx\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.056488 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/c82772b6-9737-45ee-998a-e897086c03b5-etc-ovs\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.056542 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5ql2\" (UniqueName: \"kubernetes.io/projected/b80cb988-de2a-4e65-9161-a0af0561c754-kube-api-access-h5ql2\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.056579 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/c82772b6-9737-45ee-998a-e897086c03b5-var-lib\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.056607 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/c82772b6-9737-45ee-998a-e897086c03b5-var-log\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.056635 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/b80cb988-de2a-4e65-9161-a0af0561c754-ovn-controller-tls-certs\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.056660 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b80cb988-de2a-4e65-9161-a0af0561c754-var-run\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.056707 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b80cb988-de2a-4e65-9161-a0af0561c754-combined-ca-bundle\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.056748 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c82772b6-9737-45ee-998a-e897086c03b5-var-run\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.056778 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b80cb988-de2a-4e65-9161-a0af0561c754-var-log-ovn\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.056799 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b80cb988-de2a-4e65-9161-a0af0561c754-var-run-ovn\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.056824 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b80cb988-de2a-4e65-9161-a0af0561c754-scripts\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.056845 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c82772b6-9737-45ee-998a-e897086c03b5-scripts\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.059251 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c82772b6-9737-45ee-998a-e897086c03b5-scripts\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.059965 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/c82772b6-9737-45ee-998a-e897086c03b5-etc-ovs\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.060345 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/c82772b6-9737-45ee-998a-e897086c03b5-var-lib\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.060478 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/c82772b6-9737-45ee-998a-e897086c03b5-var-log\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.061113 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b80cb988-de2a-4e65-9161-a0af0561c754-var-run\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.061146 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c82772b6-9737-45ee-998a-e897086c03b5-var-run\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.061188 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b80cb988-de2a-4e65-9161-a0af0561c754-var-run-ovn\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.061228 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b80cb988-de2a-4e65-9161-a0af0561c754-var-log-ovn\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.062929 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b80cb988-de2a-4e65-9161-a0af0561c754-scripts\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.064250 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/b80cb988-de2a-4e65-9161-a0af0561c754-ovn-controller-tls-certs\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.066568 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b80cb988-de2a-4e65-9161-a0af0561c754-combined-ca-bundle\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.075223 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hn4rx\" (UniqueName: \"kubernetes.io/projected/c82772b6-9737-45ee-998a-e897086c03b5-kube-api-access-hn4rx\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.079505 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5ql2\" (UniqueName: \"kubernetes.io/projected/b80cb988-de2a-4e65-9161-a0af0561c754-kube-api-access-h5ql2\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.171128 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-mkpng" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.186655 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.731534 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.732855 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.739938 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.740367 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.740666 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-lz8q6" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.740828 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.740910 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.745964 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.869209 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d681442e-6bdb-46e4-9b49-6466e5e036a0-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.869271 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.869294 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d681442e-6bdb-46e4-9b49-6466e5e036a0-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.869311 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d681442e-6bdb-46e4-9b49-6466e5e036a0-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.869336 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d681442e-6bdb-46e4-9b49-6466e5e036a0-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.869377 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d681442e-6bdb-46e4-9b49-6466e5e036a0-config\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.869391 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjb6g\" (UniqueName: \"kubernetes.io/projected/d681442e-6bdb-46e4-9b49-6466e5e036a0-kube-api-access-bjb6g\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.869444 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d681442e-6bdb-46e4-9b49-6466e5e036a0-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.971283 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d681442e-6bdb-46e4-9b49-6466e5e036a0-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.971340 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d681442e-6bdb-46e4-9b49-6466e5e036a0-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.971376 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.971391 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d681442e-6bdb-46e4-9b49-6466e5e036a0-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.971410 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d681442e-6bdb-46e4-9b49-6466e5e036a0-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.971457 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d681442e-6bdb-46e4-9b49-6466e5e036a0-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.971495 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d681442e-6bdb-46e4-9b49-6466e5e036a0-config\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.971515 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjb6g\" (UniqueName: \"kubernetes.io/projected/d681442e-6bdb-46e4-9b49-6466e5e036a0-kube-api-access-bjb6g\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.973073 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d681442e-6bdb-46e4-9b49-6466e5e036a0-config\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.973126 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.973525 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d681442e-6bdb-46e4-9b49-6466e5e036a0-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.974618 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d681442e-6bdb-46e4-9b49-6466e5e036a0-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.979327 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d681442e-6bdb-46e4-9b49-6466e5e036a0-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.981747 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d681442e-6bdb-46e4-9b49-6466e5e036a0-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.982318 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d681442e-6bdb-46e4-9b49-6466e5e036a0-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:53 crc kubenswrapper[4948]: I0220 08:20:53.000910 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjb6g\" (UniqueName: \"kubernetes.io/projected/d681442e-6bdb-46e4-9b49-6466e5e036a0-kube-api-access-bjb6g\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:53 crc kubenswrapper[4948]: I0220 08:20:53.015098 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:53 crc kubenswrapper[4948]: I0220 08:20:53.051331 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:55 crc kubenswrapper[4948]: I0220 08:20:55.849701 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 20 08:20:55 crc kubenswrapper[4948]: I0220 08:20:55.851516 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:55 crc kubenswrapper[4948]: I0220 08:20:55.854528 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Feb 20 08:20:55 crc kubenswrapper[4948]: I0220 08:20:55.855070 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Feb 20 08:20:55 crc kubenswrapper[4948]: I0220 08:20:55.855120 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Feb 20 08:20:55 crc kubenswrapper[4948]: I0220 08:20:55.855159 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-2kzt2" Feb 20 08:20:55 crc kubenswrapper[4948]: I0220 08:20:55.861662 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 20 08:20:55 crc kubenswrapper[4948]: I0220 08:20:55.929682 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf8w4\" (UniqueName: \"kubernetes.io/projected/4b04e19d-cd3d-474d-884d-df12d0e1fffd-kube-api-access-wf8w4\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:55 crc kubenswrapper[4948]: I0220 08:20:55.929748 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4b04e19d-cd3d-474d-884d-df12d0e1fffd-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:55 crc kubenswrapper[4948]: I0220 08:20:55.929769 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:55 crc kubenswrapper[4948]: I0220 08:20:55.929791 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b04e19d-cd3d-474d-884d-df12d0e1fffd-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:55 crc kubenswrapper[4948]: I0220 08:20:55.929818 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b04e19d-cd3d-474d-884d-df12d0e1fffd-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:55 crc kubenswrapper[4948]: I0220 08:20:55.930049 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4b04e19d-cd3d-474d-884d-df12d0e1fffd-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:55 crc kubenswrapper[4948]: I0220 08:20:55.930088 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b04e19d-cd3d-474d-884d-df12d0e1fffd-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:55 crc kubenswrapper[4948]: I0220 08:20:55.931077 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b04e19d-cd3d-474d-884d-df12d0e1fffd-config\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.033415 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf8w4\" (UniqueName: \"kubernetes.io/projected/4b04e19d-cd3d-474d-884d-df12d0e1fffd-kube-api-access-wf8w4\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.033657 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4b04e19d-cd3d-474d-884d-df12d0e1fffd-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.033688 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.033761 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b04e19d-cd3d-474d-884d-df12d0e1fffd-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.033827 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b04e19d-cd3d-474d-884d-df12d0e1fffd-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.033872 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4b04e19d-cd3d-474d-884d-df12d0e1fffd-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.033918 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b04e19d-cd3d-474d-884d-df12d0e1fffd-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.033951 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b04e19d-cd3d-474d-884d-df12d0e1fffd-config\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.034289 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.034907 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b04e19d-cd3d-474d-884d-df12d0e1fffd-config\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.035050 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4b04e19d-cd3d-474d-884d-df12d0e1fffd-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.035079 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4b04e19d-cd3d-474d-884d-df12d0e1fffd-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.041071 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b04e19d-cd3d-474d-884d-df12d0e1fffd-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.041247 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b04e19d-cd3d-474d-884d-df12d0e1fffd-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.048637 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b04e19d-cd3d-474d-884d-df12d0e1fffd-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.053326 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf8w4\" (UniqueName: \"kubernetes.io/projected/4b04e19d-cd3d-474d-884d-df12d0e1fffd-kube-api-access-wf8w4\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.065128 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.180926 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 20 08:21:01 crc kubenswrapper[4948]: E0220 08:21:01.610934 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 20 08:21:01 crc kubenswrapper[4948]: E0220 08:21:01.612166 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hdnm9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-mkkzf_openstack(5686ae39-4e44-4303-9b44-ad4b9ceae1ba): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 08:21:01 crc kubenswrapper[4948]: E0220 08:21:01.613284 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" podUID="5686ae39-4e44-4303-9b44-ad4b9ceae1ba" Feb 20 08:21:02 crc kubenswrapper[4948]: E0220 08:21:02.179064 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" podUID="5686ae39-4e44-4303-9b44-ad4b9ceae1ba" Feb 20 08:21:02 crc kubenswrapper[4948]: E0220 08:21:02.796067 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 20 08:21:02 crc kubenswrapper[4948]: E0220 08:21:02.796880 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pds7s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-zt6gg_openstack(977992e9-30ca-4420-98ae-92f91545cd10): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 08:21:02 crc kubenswrapper[4948]: E0220 08:21:02.798241 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" podUID="977992e9-30ca-4420-98ae-92f91545cd10" Feb 20 08:21:02 crc kubenswrapper[4948]: E0220 08:21:02.832203 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 20 08:21:02 crc kubenswrapper[4948]: E0220 08:21:02.832341 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ffpsp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-cqgz2_openstack(05b1d586-95db-4795-a032-f22ba90e6c16): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 08:21:02 crc kubenswrapper[4948]: E0220 08:21:02.833931 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-cqgz2" podUID="05b1d586-95db-4795-a032-f22ba90e6c16" Feb 20 08:21:02 crc kubenswrapper[4948]: E0220 08:21:02.876455 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 20 08:21:02 crc kubenswrapper[4948]: E0220 08:21:02.876617 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7d754,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-9dfhs_openstack(1224cf30-813c-48f2-92c9-fabe78bbc6bc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 08:21:02 crc kubenswrapper[4948]: E0220 08:21:02.880070 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-9dfhs" podUID="1224cf30-813c-48f2-92c9-fabe78bbc6bc" Feb 20 08:21:03 crc kubenswrapper[4948]: E0220 08:21:03.186267 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" podUID="977992e9-30ca-4420-98ae-92f91545cd10" Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.333268 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 20 08:21:03 crc kubenswrapper[4948]: W0220 08:21:03.348135 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8e994489_a2cf_4f6d_a00c_98f627ba0e5f.slice/crio-1e7df2ea4e6039fb737b56f6e8e97afb2ce13f39013495a5535bd3b4cf07ba35 WatchSource:0}: Error finding container 1e7df2ea4e6039fb737b56f6e8e97afb2ce13f39013495a5535bd3b4cf07ba35: Status 404 returned error can't find the container with id 1e7df2ea4e6039fb737b56f6e8e97afb2ce13f39013495a5535bd3b4cf07ba35 Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.359579 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 20 08:21:03 crc kubenswrapper[4948]: W0220 08:21:03.371115 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod790894c9_053c_497c_955d_ce7519111dd6.slice/crio-39af94a050f39ff7ebfa915fc855416f40bbca9d8ae91ba23049ec9be5e37b76 WatchSource:0}: Error finding container 39af94a050f39ff7ebfa915fc855416f40bbca9d8ae91ba23049ec9be5e37b76: Status 404 returned error can't find the container with id 39af94a050f39ff7ebfa915fc855416f40bbca9d8ae91ba23049ec9be5e37b76 Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.513939 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-cqgz2" Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.529786 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-9dfhs" Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.584619 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05b1d586-95db-4795-a032-f22ba90e6c16-dns-svc\") pod \"05b1d586-95db-4795-a032-f22ba90e6c16\" (UID: \"05b1d586-95db-4795-a032-f22ba90e6c16\") " Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.584710 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05b1d586-95db-4795-a032-f22ba90e6c16-config\") pod \"05b1d586-95db-4795-a032-f22ba90e6c16\" (UID: \"05b1d586-95db-4795-a032-f22ba90e6c16\") " Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.584791 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffpsp\" (UniqueName: \"kubernetes.io/projected/05b1d586-95db-4795-a032-f22ba90e6c16-kube-api-access-ffpsp\") pod \"05b1d586-95db-4795-a032-f22ba90e6c16\" (UID: \"05b1d586-95db-4795-a032-f22ba90e6c16\") " Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.585322 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05b1d586-95db-4795-a032-f22ba90e6c16-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "05b1d586-95db-4795-a032-f22ba90e6c16" (UID: "05b1d586-95db-4795-a032-f22ba90e6c16"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.585512 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05b1d586-95db-4795-a032-f22ba90e6c16-config" (OuterVolumeSpecName: "config") pod "05b1d586-95db-4795-a032-f22ba90e6c16" (UID: "05b1d586-95db-4795-a032-f22ba90e6c16"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.588863 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05b1d586-95db-4795-a032-f22ba90e6c16-kube-api-access-ffpsp" (OuterVolumeSpecName: "kube-api-access-ffpsp") pod "05b1d586-95db-4795-a032-f22ba90e6c16" (UID: "05b1d586-95db-4795-a032-f22ba90e6c16"). InnerVolumeSpecName "kube-api-access-ffpsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.634104 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rt8tp"] Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.649591 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.658492 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.664614 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-mkpng"] Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.670588 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.685865 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1224cf30-813c-48f2-92c9-fabe78bbc6bc-config\") pod \"1224cf30-813c-48f2-92c9-fabe78bbc6bc\" (UID: \"1224cf30-813c-48f2-92c9-fabe78bbc6bc\") " Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.685901 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7d754\" (UniqueName: \"kubernetes.io/projected/1224cf30-813c-48f2-92c9-fabe78bbc6bc-kube-api-access-7d754\") pod \"1224cf30-813c-48f2-92c9-fabe78bbc6bc\" (UID: \"1224cf30-813c-48f2-92c9-fabe78bbc6bc\") " Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.686292 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffpsp\" (UniqueName: \"kubernetes.io/projected/05b1d586-95db-4795-a032-f22ba90e6c16-kube-api-access-ffpsp\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.686311 4948 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05b1d586-95db-4795-a032-f22ba90e6c16-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.686322 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05b1d586-95db-4795-a032-f22ba90e6c16-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.687184 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1224cf30-813c-48f2-92c9-fabe78bbc6bc-config" (OuterVolumeSpecName: "config") pod "1224cf30-813c-48f2-92c9-fabe78bbc6bc" (UID: "1224cf30-813c-48f2-92c9-fabe78bbc6bc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.734937 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1224cf30-813c-48f2-92c9-fabe78bbc6bc-kube-api-access-7d754" (OuterVolumeSpecName: "kube-api-access-7d754") pod "1224cf30-813c-48f2-92c9-fabe78bbc6bc" (UID: "1224cf30-813c-48f2-92c9-fabe78bbc6bc"). InnerVolumeSpecName "kube-api-access-7d754". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.786620 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-q4bv4"] Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.787833 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1224cf30-813c-48f2-92c9-fabe78bbc6bc-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.787870 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7d754\" (UniqueName: \"kubernetes.io/projected/1224cf30-813c-48f2-92c9-fabe78bbc6bc-kube-api-access-7d754\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:03 crc kubenswrapper[4948]: W0220 08:21:03.790500 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0dfbb200_e444_4895_a410_e51cbb1fc112.slice/crio-8d55ac91670695a37a460e364e6527e261fc751cd8cb9dd56d60bf1aa04e7514 WatchSource:0}: Error finding container 8d55ac91670695a37a460e364e6527e261fc751cd8cb9dd56d60bf1aa04e7514: Status 404 returned error can't find the container with id 8d55ac91670695a37a460e364e6527e261fc751cd8cb9dd56d60bf1aa04e7514 Feb 20 08:21:03 crc kubenswrapper[4948]: W0220 08:21:03.801444 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc82772b6_9737_45ee_998a_e897086c03b5.slice/crio-ec18154bf98d02691b4396af19785093698c4033031de11ba963723e7ece9f00 WatchSource:0}: Error finding container ec18154bf98d02691b4396af19785093698c4033031de11ba963723e7ece9f00: Status 404 returned error can't find the container with id ec18154bf98d02691b4396af19785093698c4033031de11ba963723e7ece9f00 Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.190623 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-9dfhs" event={"ID":"1224cf30-813c-48f2-92c9-fabe78bbc6bc","Type":"ContainerDied","Data":"459736be569ebf9acfc7e517d56782883ab07764b913b8c4d305735259e95b55"} Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.190723 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-9dfhs" Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.192885 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"0dfbb200-e444-4895-a410-e51cbb1fc112","Type":"ContainerStarted","Data":"8d55ac91670695a37a460e364e6527e261fc751cd8cb9dd56d60bf1aa04e7514"} Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.197286 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"73b84bb7-f594-4823-ac03-40fdac6ee177","Type":"ContainerStarted","Data":"d4b1f06d6e332e362b758285a5271800ff6a5cc95aee58142ef4093fc832e364"} Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.207417 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"790894c9-053c-497c-955d-ce7519111dd6","Type":"ContainerStarted","Data":"39af94a050f39ff7ebfa915fc855416f40bbca9d8ae91ba23049ec9be5e37b76"} Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.208741 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d02114ae-21f2-4cc5-9bca-9505c25eaf5f","Type":"ContainerStarted","Data":"88f7820770eaae07d0058f62cddfa17073943d7aca489b897da76d305b4a45a0"} Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.210430 4948 generic.go:334] "Generic (PLEG): container finished" podID="93f24735-d27d-4867-b9f7-cce47e7f376f" containerID="b689fa43e3b5c02fa357d00443dc58a7d6fa6d71f6692b435028801366e8a5e6" exitCode=0 Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.210489 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rt8tp" event={"ID":"93f24735-d27d-4867-b9f7-cce47e7f376f","Type":"ContainerDied","Data":"b689fa43e3b5c02fa357d00443dc58a7d6fa6d71f6692b435028801366e8a5e6"} Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.210511 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rt8tp" event={"ID":"93f24735-d27d-4867-b9f7-cce47e7f376f","Type":"ContainerStarted","Data":"104e51617975f8f0d6bf864e106e2486e0a18298cc9a24f078002083a6d423b2"} Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.212267 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-mkpng" event={"ID":"b80cb988-de2a-4e65-9161-a0af0561c754","Type":"ContainerStarted","Data":"128bc4be043ddeeae350cc9c95a3938b26679c4935fdd074d497449c5afcb7a9"} Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.213323 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"8e994489-a2cf-4f6d-a00c-98f627ba0e5f","Type":"ContainerStarted","Data":"1e7df2ea4e6039fb737b56f6e8e97afb2ce13f39013495a5535bd3b4cf07ba35"} Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.217199 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-cqgz2" event={"ID":"05b1d586-95db-4795-a032-f22ba90e6c16","Type":"ContainerDied","Data":"4e5e167063afcb1f5336b93cee715fd831f959a4a5fb12e9e562274ec18ad672"} Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.217322 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-cqgz2" Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.247376 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6e59b63d-0c56-488c-87cb-348af87058c6","Type":"ContainerStarted","Data":"0391c43404c3571e63693b407b5d3441d92363fdd709b3395357f5dcce9c9f30"} Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.249568 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-q4bv4" event={"ID":"c82772b6-9737-45ee-998a-e897086c03b5","Type":"ContainerStarted","Data":"ec18154bf98d02691b4396af19785093698c4033031de11ba963723e7ece9f00"} Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.262656 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9dfhs"] Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.265640 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9dfhs"] Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.292373 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-cqgz2"] Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.297273 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-cqgz2"] Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.350716 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 20 08:21:04 crc kubenswrapper[4948]: W0220 08:21:04.357245 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd681442e_6bdb_46e4_9b49_6466e5e036a0.slice/crio-d8ed751d1b95575a7a1d2d4ba48699ab1ced21c222a4164eb4cdb6f411745130 WatchSource:0}: Error finding container d8ed751d1b95575a7a1d2d4ba48699ab1ced21c222a4164eb4cdb6f411745130: Status 404 returned error can't find the container with id d8ed751d1b95575a7a1d2d4ba48699ab1ced21c222a4164eb4cdb6f411745130 Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.585935 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 20 08:21:04 crc kubenswrapper[4948]: W0220 08:21:04.923943 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b04e19d_cd3d_474d_884d_df12d0e1fffd.slice/crio-660ea7b70688ba8b37a938c6b7d318fb1a1c77090b2c1e58092a4070f0016d7c WatchSource:0}: Error finding container 660ea7b70688ba8b37a938c6b7d318fb1a1c77090b2c1e58092a4070f0016d7c: Status 404 returned error can't find the container with id 660ea7b70688ba8b37a938c6b7d318fb1a1c77090b2c1e58092a4070f0016d7c Feb 20 08:21:05 crc kubenswrapper[4948]: I0220 08:21:05.257388 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"790894c9-053c-497c-955d-ce7519111dd6","Type":"ContainerStarted","Data":"77b026b61409d73dee6a89b7f49bf1a86167df7972db0e3eb3f4cbb10df48242"} Feb 20 08:21:05 crc kubenswrapper[4948]: I0220 08:21:05.260137 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"d681442e-6bdb-46e4-9b49-6466e5e036a0","Type":"ContainerStarted","Data":"d8ed751d1b95575a7a1d2d4ba48699ab1ced21c222a4164eb4cdb6f411745130"} Feb 20 08:21:05 crc kubenswrapper[4948]: I0220 08:21:05.262086 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"4b04e19d-cd3d-474d-884d-df12d0e1fffd","Type":"ContainerStarted","Data":"660ea7b70688ba8b37a938c6b7d318fb1a1c77090b2c1e58092a4070f0016d7c"} Feb 20 08:21:06 crc kubenswrapper[4948]: I0220 08:21:05.737536 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05b1d586-95db-4795-a032-f22ba90e6c16" path="/var/lib/kubelet/pods/05b1d586-95db-4795-a032-f22ba90e6c16/volumes" Feb 20 08:21:06 crc kubenswrapper[4948]: I0220 08:21:05.738726 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1224cf30-813c-48f2-92c9-fabe78bbc6bc" path="/var/lib/kubelet/pods/1224cf30-813c-48f2-92c9-fabe78bbc6bc/volumes" Feb 20 08:21:08 crc kubenswrapper[4948]: I0220 08:21:08.025005 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:21:08 crc kubenswrapper[4948]: I0220 08:21:08.025498 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.365859 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6e59b63d-0c56-488c-87cb-348af87058c6","Type":"ContainerStarted","Data":"52aab5e3e31e0e0c819bc56b6e3130d428a9243c172c9a352f49b7f3ba0db78d"} Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.368167 4948 generic.go:334] "Generic (PLEG): container finished" podID="c82772b6-9737-45ee-998a-e897086c03b5" containerID="ff1b1e67637ea3148c6b86c285290aa88c29f45d19a025c92565c44923597b25" exitCode=0 Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.368222 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-q4bv4" event={"ID":"c82772b6-9737-45ee-998a-e897086c03b5","Type":"ContainerDied","Data":"ff1b1e67637ea3148c6b86c285290aa88c29f45d19a025c92565c44923597b25"} Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.372242 4948 generic.go:334] "Generic (PLEG): container finished" podID="93f24735-d27d-4867-b9f7-cce47e7f376f" containerID="7f925a093fb8a5b68097ec9940401bb5483d01b61401a6a70fda1f29efe6615e" exitCode=0 Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.372280 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rt8tp" event={"ID":"93f24735-d27d-4867-b9f7-cce47e7f376f","Type":"ContainerDied","Data":"7f925a093fb8a5b68097ec9940401bb5483d01b61401a6a70fda1f29efe6615e"} Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.375792 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d02114ae-21f2-4cc5-9bca-9505c25eaf5f","Type":"ContainerStarted","Data":"3ef0ae79740a0d142692c9fc559c45d8a8e3955ca475b2d1eb3cc2ae8aea4eac"} Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.377904 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"8e994489-a2cf-4f6d-a00c-98f627ba0e5f","Type":"ContainerStarted","Data":"6209e5b7193825550a19ad6f82db51db28926334103fbbb1ba1accdf4480217a"} Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.378350 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.385275 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"4b04e19d-cd3d-474d-884d-df12d0e1fffd","Type":"ContainerStarted","Data":"2982b3c30a182ecd9b19a054cd88fa344147eba3d2ef83da730f1b6236bef02e"} Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.387960 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"0dfbb200-e444-4895-a410-e51cbb1fc112","Type":"ContainerStarted","Data":"20c91950ce09f61ade4d935382ea88014ac2e6914fdd0f747d53e6577cf780a0"} Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.388842 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.391303 4948 generic.go:334] "Generic (PLEG): container finished" podID="5686ae39-4e44-4303-9b44-ad4b9ceae1ba" containerID="5a5bfff9ba7b803c56d99738536efa4fb27865dfba1fd58a67dc23b1d4bef9ed" exitCode=0 Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.391380 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" event={"ID":"5686ae39-4e44-4303-9b44-ad4b9ceae1ba","Type":"ContainerDied","Data":"5a5bfff9ba7b803c56d99738536efa4fb27865dfba1fd58a67dc23b1d4bef9ed"} Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.404309 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-mkpng" event={"ID":"b80cb988-de2a-4e65-9161-a0af0561c754","Type":"ContainerStarted","Data":"99d11716eb57476f8ffaf8e1e981cf08ec5515e1fcf0b7d0d1ca6915191b7c59"} Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.404472 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-mkpng" Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.409795 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"d681442e-6bdb-46e4-9b49-6466e5e036a0","Type":"ContainerStarted","Data":"154b3b91cde6adf519be9c377c4fa145eb9c9b5eb16cfea11a3489cc91629ba5"} Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.429330 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=19.895896955 podStartE2EDuration="30.429316495s" podCreationTimestamp="2026-02-20 08:20:46 +0000 UTC" firstStartedPulling="2026-02-20 08:21:03.357149088 +0000 UTC m=+912.331643928" lastFinishedPulling="2026-02-20 08:21:13.890568638 +0000 UTC m=+922.865063468" observedRunningTime="2026-02-20 08:21:16.428549546 +0000 UTC m=+925.403044366" watchObservedRunningTime="2026-02-20 08:21:16.429316495 +0000 UTC m=+925.403811305" Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.495931 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=15.880916618 podStartE2EDuration="27.495908993s" podCreationTimestamp="2026-02-20 08:20:49 +0000 UTC" firstStartedPulling="2026-02-20 08:21:03.795894026 +0000 UTC m=+912.770388846" lastFinishedPulling="2026-02-20 08:21:15.410886401 +0000 UTC m=+924.385381221" observedRunningTime="2026-02-20 08:21:16.481500036 +0000 UTC m=+925.455994856" watchObservedRunningTime="2026-02-20 08:21:16.495908993 +0000 UTC m=+925.470403813" Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.502306 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-mkpng" podStartSLOduration=14.949192415 podStartE2EDuration="25.502290311s" podCreationTimestamp="2026-02-20 08:20:51 +0000 UTC" firstStartedPulling="2026-02-20 08:21:03.76614449 +0000 UTC m=+912.740639310" lastFinishedPulling="2026-02-20 08:21:14.319242386 +0000 UTC m=+923.293737206" observedRunningTime="2026-02-20 08:21:16.497633985 +0000 UTC m=+925.472128815" watchObservedRunningTime="2026-02-20 08:21:16.502290311 +0000 UTC m=+925.476785131" Feb 20 08:21:17 crc kubenswrapper[4948]: I0220 08:21:17.420666 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" event={"ID":"5686ae39-4e44-4303-9b44-ad4b9ceae1ba","Type":"ContainerStarted","Data":"f8a24951a6fecaa760fab34934fbaaebc5ce586261d093c015219f82cccd0422"} Feb 20 08:21:17 crc kubenswrapper[4948]: I0220 08:21:17.421379 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" Feb 20 08:21:17 crc kubenswrapper[4948]: I0220 08:21:17.425398 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-q4bv4" event={"ID":"c82772b6-9737-45ee-998a-e897086c03b5","Type":"ContainerStarted","Data":"52a4b050691a7d6574abd457527106d71c476df878106b33767014ec3321039e"} Feb 20 08:21:18 crc kubenswrapper[4948]: I0220 08:21:18.435292 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"4b04e19d-cd3d-474d-884d-df12d0e1fffd","Type":"ContainerStarted","Data":"2e87e5a50286fc9148333fb04efebbd5fb4d85acaf989d4e9a4d48c3df249122"} Feb 20 08:21:18 crc kubenswrapper[4948]: I0220 08:21:18.437569 4948 generic.go:334] "Generic (PLEG): container finished" podID="977992e9-30ca-4420-98ae-92f91545cd10" containerID="05ccd3525150ab098e3561398cfda7a7a04a42e4d12ac43eacc78bfa91ca9099" exitCode=0 Feb 20 08:21:18 crc kubenswrapper[4948]: I0220 08:21:18.437647 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" event={"ID":"977992e9-30ca-4420-98ae-92f91545cd10","Type":"ContainerDied","Data":"05ccd3525150ab098e3561398cfda7a7a04a42e4d12ac43eacc78bfa91ca9099"} Feb 20 08:21:18 crc kubenswrapper[4948]: I0220 08:21:18.440734 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-q4bv4" event={"ID":"c82772b6-9737-45ee-998a-e897086c03b5","Type":"ContainerStarted","Data":"45d366f5451bce0eea4471ff0fa712033b40558f9b4867b32e46492ba2b78eaa"} Feb 20 08:21:18 crc kubenswrapper[4948]: I0220 08:21:18.440910 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:21:18 crc kubenswrapper[4948]: I0220 08:21:18.444561 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rt8tp" event={"ID":"93f24735-d27d-4867-b9f7-cce47e7f376f","Type":"ContainerStarted","Data":"fbf3b4c07c31284ef058431c15f6fb31624a27a446740ecaf65fcf08b863a6e3"} Feb 20 08:21:18 crc kubenswrapper[4948]: I0220 08:21:18.447952 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"d681442e-6bdb-46e4-9b49-6466e5e036a0","Type":"ContainerStarted","Data":"c93fa988182cb818cfb4eb867eabf542e0d664bbe1a2420a5257aa6e4180960f"} Feb 20 08:21:18 crc kubenswrapper[4948]: I0220 08:21:18.483150 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=11.916423383 podStartE2EDuration="24.48312883s" podCreationTimestamp="2026-02-20 08:20:54 +0000 UTC" firstStartedPulling="2026-02-20 08:21:04.935089118 +0000 UTC m=+913.909583978" lastFinishedPulling="2026-02-20 08:21:17.501794605 +0000 UTC m=+926.476289425" observedRunningTime="2026-02-20 08:21:18.481406507 +0000 UTC m=+927.455901327" watchObservedRunningTime="2026-02-20 08:21:18.48312883 +0000 UTC m=+927.457623660" Feb 20 08:21:18 crc kubenswrapper[4948]: I0220 08:21:18.484382 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" podStartSLOduration=5.125547855 podStartE2EDuration="36.484372111s" podCreationTimestamp="2026-02-20 08:20:42 +0000 UTC" firstStartedPulling="2026-02-20 08:20:43.884879158 +0000 UTC m=+892.859373978" lastFinishedPulling="2026-02-20 08:21:15.243703394 +0000 UTC m=+924.218198234" observedRunningTime="2026-02-20 08:21:17.441761979 +0000 UTC m=+926.416256809" watchObservedRunningTime="2026-02-20 08:21:18.484372111 +0000 UTC m=+927.458866941" Feb 20 08:21:18 crc kubenswrapper[4948]: I0220 08:21:18.539493 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-q4bv4" podStartSLOduration=17.454977274 podStartE2EDuration="27.539476124s" podCreationTimestamp="2026-02-20 08:20:51 +0000 UTC" firstStartedPulling="2026-02-20 08:21:03.804730965 +0000 UTC m=+912.779225785" lastFinishedPulling="2026-02-20 08:21:13.889229805 +0000 UTC m=+922.863724635" observedRunningTime="2026-02-20 08:21:18.535343732 +0000 UTC m=+927.509838552" watchObservedRunningTime="2026-02-20 08:21:18.539476124 +0000 UTC m=+927.513970944" Feb 20 08:21:18 crc kubenswrapper[4948]: I0220 08:21:18.601658 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=14.463602167 podStartE2EDuration="27.601640393s" podCreationTimestamp="2026-02-20 08:20:51 +0000 UTC" firstStartedPulling="2026-02-20 08:21:04.36380509 +0000 UTC m=+913.338299920" lastFinishedPulling="2026-02-20 08:21:17.501843316 +0000 UTC m=+926.476338146" observedRunningTime="2026-02-20 08:21:18.573856265 +0000 UTC m=+927.548351075" watchObservedRunningTime="2026-02-20 08:21:18.601640393 +0000 UTC m=+927.576135213" Feb 20 08:21:18 crc kubenswrapper[4948]: I0220 08:21:18.703157 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rt8tp" podStartSLOduration=20.417332932 podStartE2EDuration="33.703139445s" podCreationTimestamp="2026-02-20 08:20:45 +0000 UTC" firstStartedPulling="2026-02-20 08:21:04.213406028 +0000 UTC m=+913.187900858" lastFinishedPulling="2026-02-20 08:21:17.499212551 +0000 UTC m=+926.473707371" observedRunningTime="2026-02-20 08:21:18.692630885 +0000 UTC m=+927.667125695" watchObservedRunningTime="2026-02-20 08:21:18.703139445 +0000 UTC m=+927.677634265" Feb 20 08:21:19 crc kubenswrapper[4948]: I0220 08:21:19.459460 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" event={"ID":"977992e9-30ca-4420-98ae-92f91545cd10","Type":"ContainerStarted","Data":"2fa62b0646099a9e4d67d43a31c1228c43fee77b21220b1456b925205d581d9a"} Feb 20 08:21:19 crc kubenswrapper[4948]: I0220 08:21:19.460885 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" Feb 20 08:21:19 crc kubenswrapper[4948]: I0220 08:21:19.461482 4948 generic.go:334] "Generic (PLEG): container finished" podID="6e59b63d-0c56-488c-87cb-348af87058c6" containerID="52aab5e3e31e0e0c819bc56b6e3130d428a9243c172c9a352f49b7f3ba0db78d" exitCode=0 Feb 20 08:21:19 crc kubenswrapper[4948]: I0220 08:21:19.461575 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6e59b63d-0c56-488c-87cb-348af87058c6","Type":"ContainerDied","Data":"52aab5e3e31e0e0c819bc56b6e3130d428a9243c172c9a352f49b7f3ba0db78d"} Feb 20 08:21:19 crc kubenswrapper[4948]: I0220 08:21:19.462928 4948 generic.go:334] "Generic (PLEG): container finished" podID="d02114ae-21f2-4cc5-9bca-9505c25eaf5f" containerID="3ef0ae79740a0d142692c9fc559c45d8a8e3955ca475b2d1eb3cc2ae8aea4eac" exitCode=0 Feb 20 08:21:19 crc kubenswrapper[4948]: I0220 08:21:19.463141 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d02114ae-21f2-4cc5-9bca-9505c25eaf5f","Type":"ContainerDied","Data":"3ef0ae79740a0d142692c9fc559c45d8a8e3955ca475b2d1eb3cc2ae8aea4eac"} Feb 20 08:21:19 crc kubenswrapper[4948]: I0220 08:21:19.464834 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:21:19 crc kubenswrapper[4948]: I0220 08:21:19.494654 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" podStartSLOduration=-9223371999.360136 podStartE2EDuration="37.494639802s" podCreationTimestamp="2026-02-20 08:20:42 +0000 UTC" firstStartedPulling="2026-02-20 08:20:43.472030076 +0000 UTC m=+892.446524896" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:21:19.493017722 +0000 UTC m=+928.467512582" watchObservedRunningTime="2026-02-20 08:21:19.494639802 +0000 UTC m=+928.469134622" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.052900 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.104227 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.182126 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.225918 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.477744 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d02114ae-21f2-4cc5-9bca-9505c25eaf5f","Type":"ContainerStarted","Data":"71988e03aff6de789cff2b8a78a3ac5d432604d0dc5f50a7f67a3f2ee9dd935d"} Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.482415 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6e59b63d-0c56-488c-87cb-348af87058c6","Type":"ContainerStarted","Data":"c46f534c65704b9778549e3e9e455d8942dd90bd2303681319fcb6071803ed28"} Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.483573 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.483609 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.511433 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=25.376815015 podStartE2EDuration="35.511417725s" podCreationTimestamp="2026-02-20 08:20:45 +0000 UTC" firstStartedPulling="2026-02-20 08:21:03.754600284 +0000 UTC m=+912.729095104" lastFinishedPulling="2026-02-20 08:21:13.889202994 +0000 UTC m=+922.863697814" observedRunningTime="2026-02-20 08:21:20.509831225 +0000 UTC m=+929.484326045" watchObservedRunningTime="2026-02-20 08:21:20.511417725 +0000 UTC m=+929.485912545" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.544375 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=25.895116594 podStartE2EDuration="36.54435095s" podCreationTimestamp="2026-02-20 08:20:44 +0000 UTC" firstStartedPulling="2026-02-20 08:21:03.754561803 +0000 UTC m=+912.729056623" lastFinishedPulling="2026-02-20 08:21:14.403796149 +0000 UTC m=+923.378290979" observedRunningTime="2026-02-20 08:21:20.537819448 +0000 UTC m=+929.512314278" watchObservedRunningTime="2026-02-20 08:21:20.54435095 +0000 UTC m=+929.518845780" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.552127 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.555522 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.721913 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-mkkzf"] Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.723084 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" podUID="5686ae39-4e44-4303-9b44-ad4b9ceae1ba" containerName="dnsmasq-dns" containerID="cri-o://f8a24951a6fecaa760fab34934fbaaebc5ce586261d093c015219f82cccd0422" gracePeriod=10 Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.774601 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-mvb6f"] Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.776106 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.778073 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.781224 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-gcfn8"] Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.784095 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.785421 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.849286 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-gcfn8"] Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.858502 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1510e8b2-af6f-4641-a2d2-361830382250-config\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.858583 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cpkv\" (UniqueName: \"kubernetes.io/projected/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-kube-api-access-9cpkv\") pod \"dnsmasq-dns-7fd796d7df-mvb6f\" (UID: \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\") " pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.858602 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-mvb6f\" (UID: \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\") " pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.858626 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1510e8b2-af6f-4641-a2d2-361830382250-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.858653 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kw495\" (UniqueName: \"kubernetes.io/projected/1510e8b2-af6f-4641-a2d2-361830382250-kube-api-access-kw495\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.858673 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/1510e8b2-af6f-4641-a2d2-361830382250-ovs-rundir\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.858725 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1510e8b2-af6f-4641-a2d2-361830382250-combined-ca-bundle\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.858739 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/1510e8b2-af6f-4641-a2d2-361830382250-ovn-rundir\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.858758 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-config\") pod \"dnsmasq-dns-7fd796d7df-mvb6f\" (UID: \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\") " pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.858784 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-mvb6f\" (UID: \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\") " pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.878432 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-mvb6f"] Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.914324 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zt6gg"] Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.949310 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-hmhzl"] Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.950550 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.960449 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.961490 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-mvb6f\" (UID: \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\") " pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.961543 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1510e8b2-af6f-4641-a2d2-361830382250-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.961575 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kw495\" (UniqueName: \"kubernetes.io/projected/1510e8b2-af6f-4641-a2d2-361830382250-kube-api-access-kw495\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.961603 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/1510e8b2-af6f-4641-a2d2-361830382250-ovs-rundir\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.961649 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1510e8b2-af6f-4641-a2d2-361830382250-combined-ca-bundle\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.961664 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/1510e8b2-af6f-4641-a2d2-361830382250-ovn-rundir\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.961687 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-config\") pod \"dnsmasq-dns-7fd796d7df-mvb6f\" (UID: \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\") " pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.961717 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-mvb6f\" (UID: \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\") " pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.961768 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1510e8b2-af6f-4641-a2d2-361830382250-config\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.961820 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cpkv\" (UniqueName: \"kubernetes.io/projected/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-kube-api-access-9cpkv\") pod \"dnsmasq-dns-7fd796d7df-mvb6f\" (UID: \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\") " pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.963220 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/1510e8b2-af6f-4641-a2d2-361830382250-ovn-rundir\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.964755 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/1510e8b2-af6f-4641-a2d2-361830382250-ovs-rundir\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.964984 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-config\") pod \"dnsmasq-dns-7fd796d7df-mvb6f\" (UID: \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\") " pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.965399 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-mvb6f\" (UID: \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\") " pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.965546 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-mvb6f\" (UID: \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\") " pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.965668 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1510e8b2-af6f-4641-a2d2-361830382250-config\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.972059 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1510e8b2-af6f-4641-a2d2-361830382250-combined-ca-bundle\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.975275 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1510e8b2-af6f-4641-a2d2-361830382250-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.976776 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-hmhzl"] Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.987115 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cpkv\" (UniqueName: \"kubernetes.io/projected/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-kube-api-access-9cpkv\") pod \"dnsmasq-dns-7fd796d7df-mvb6f\" (UID: \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\") " pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.989050 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kw495\" (UniqueName: \"kubernetes.io/projected/1510e8b2-af6f-4641-a2d2-361830382250-kube-api-access-kw495\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.004414 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.006361 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.010914 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.019195 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-bpq2x" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.019348 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.019469 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.019651 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.063841 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/eeb3782c-0a88-4d17-b481-e2aef7ea386a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.063955 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-config\") pod \"dnsmasq-dns-86db49b7ff-hmhzl\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.064035 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-hmhzl\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.064086 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-hmhzl\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.064192 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/eeb3782c-0a88-4d17-b481-e2aef7ea386a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.064227 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7h4r\" (UniqueName: \"kubernetes.io/projected/eeb3782c-0a88-4d17-b481-e2aef7ea386a-kube-api-access-c7h4r\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.064279 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eeb3782c-0a88-4d17-b481-e2aef7ea386a-scripts\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.064328 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-hmhzl\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.064468 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eeb3782c-0a88-4d17-b481-e2aef7ea386a-config\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.064490 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeb3782c-0a88-4d17-b481-e2aef7ea386a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.064506 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2twp\" (UniqueName: \"kubernetes.io/projected/33a06f41-ea8b-4138-8048-5425cd04f015-kube-api-access-c2twp\") pod \"dnsmasq-dns-86db49b7ff-hmhzl\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.064663 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/eeb3782c-0a88-4d17-b481-e2aef7ea386a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.166452 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-hmhzl\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.166497 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-hmhzl\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.166544 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/eeb3782c-0a88-4d17-b481-e2aef7ea386a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.166565 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7h4r\" (UniqueName: \"kubernetes.io/projected/eeb3782c-0a88-4d17-b481-e2aef7ea386a-kube-api-access-c7h4r\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.166599 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eeb3782c-0a88-4d17-b481-e2aef7ea386a-scripts\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.166628 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-hmhzl\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.166659 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeb3782c-0a88-4d17-b481-e2aef7ea386a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.166679 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2twp\" (UniqueName: \"kubernetes.io/projected/33a06f41-ea8b-4138-8048-5425cd04f015-kube-api-access-c2twp\") pod \"dnsmasq-dns-86db49b7ff-hmhzl\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.166700 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eeb3782c-0a88-4d17-b481-e2aef7ea386a-config\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.166733 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/eeb3782c-0a88-4d17-b481-e2aef7ea386a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.166763 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/eeb3782c-0a88-4d17-b481-e2aef7ea386a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.166823 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-config\") pod \"dnsmasq-dns-86db49b7ff-hmhzl\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.168083 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-config\") pod \"dnsmasq-dns-86db49b7ff-hmhzl\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.169060 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/eeb3782c-0a88-4d17-b481-e2aef7ea386a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.169319 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-hmhzl\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.171877 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-hmhzl\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.172580 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-hmhzl\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.173020 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eeb3782c-0a88-4d17-b481-e2aef7ea386a-scripts\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.173193 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/eeb3782c-0a88-4d17-b481-e2aef7ea386a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.173378 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeb3782c-0a88-4d17-b481-e2aef7ea386a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.173650 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/eeb3782c-0a88-4d17-b481-e2aef7ea386a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.174080 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eeb3782c-0a88-4d17-b481-e2aef7ea386a-config\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.185736 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2twp\" (UniqueName: \"kubernetes.io/projected/33a06f41-ea8b-4138-8048-5425cd04f015-kube-api-access-c2twp\") pod \"dnsmasq-dns-86db49b7ff-hmhzl\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.188430 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7h4r\" (UniqueName: \"kubernetes.io/projected/eeb3782c-0a88-4d17-b481-e2aef7ea386a-kube-api-access-c7h4r\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.191788 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.199018 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.374860 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.388830 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.494404 4948 generic.go:334] "Generic (PLEG): container finished" podID="5686ae39-4e44-4303-9b44-ad4b9ceae1ba" containerID="f8a24951a6fecaa760fab34934fbaaebc5ce586261d093c015219f82cccd0422" exitCode=0 Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.496097 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" event={"ID":"5686ae39-4e44-4303-9b44-ad4b9ceae1ba","Type":"ContainerDied","Data":"f8a24951a6fecaa760fab34934fbaaebc5ce586261d093c015219f82cccd0422"} Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.496311 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" podUID="977992e9-30ca-4420-98ae-92f91545cd10" containerName="dnsmasq-dns" containerID="cri-o://2fa62b0646099a9e4d67d43a31c1228c43fee77b21220b1456b925205d581d9a" gracePeriod=10 Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.793470 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-mvb6f"] Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.819101 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-gcfn8"] Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.845495 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-hmhzl"] Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.854647 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.143491 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.148113 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.195463 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdnm9\" (UniqueName: \"kubernetes.io/projected/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-kube-api-access-hdnm9\") pod \"5686ae39-4e44-4303-9b44-ad4b9ceae1ba\" (UID: \"5686ae39-4e44-4303-9b44-ad4b9ceae1ba\") " Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.195548 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-dns-svc\") pod \"5686ae39-4e44-4303-9b44-ad4b9ceae1ba\" (UID: \"5686ae39-4e44-4303-9b44-ad4b9ceae1ba\") " Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.195615 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-config\") pod \"5686ae39-4e44-4303-9b44-ad4b9ceae1ba\" (UID: \"5686ae39-4e44-4303-9b44-ad4b9ceae1ba\") " Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.202124 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-kube-api-access-hdnm9" (OuterVolumeSpecName: "kube-api-access-hdnm9") pod "5686ae39-4e44-4303-9b44-ad4b9ceae1ba" (UID: "5686ae39-4e44-4303-9b44-ad4b9ceae1ba"). InnerVolumeSpecName "kube-api-access-hdnm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.298314 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdnm9\" (UniqueName: \"kubernetes.io/projected/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-kube-api-access-hdnm9\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.363133 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5686ae39-4e44-4303-9b44-ad4b9ceae1ba" (UID: "5686ae39-4e44-4303-9b44-ad4b9ceae1ba"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.399796 4948 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.420489 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-config" (OuterVolumeSpecName: "config") pod "5686ae39-4e44-4303-9b44-ad4b9ceae1ba" (UID: "5686ae39-4e44-4303-9b44-ad4b9ceae1ba"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.437466 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.506952 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.519936 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-gcfn8" event={"ID":"1510e8b2-af6f-4641-a2d2-361830382250","Type":"ContainerStarted","Data":"199d2767067990d2fddd48e26caa156fee51020bbc1379e0aa8846fdfc7146d7"} Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.520000 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-gcfn8" event={"ID":"1510e8b2-af6f-4641-a2d2-361830382250","Type":"ContainerStarted","Data":"1b538d3cc5f8e21b1fd8955b6a0f0b465e33747d8eced9c6d602b588e10366c9"} Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.522045 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" event={"ID":"5686ae39-4e44-4303-9b44-ad4b9ceae1ba","Type":"ContainerDied","Data":"b4980c083e552022d3673e8c44cd0e8286ee33627cbef5030e75bd7be40133d3"} Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.522114 4948 scope.go:117] "RemoveContainer" containerID="f8a24951a6fecaa760fab34934fbaaebc5ce586261d093c015219f82cccd0422" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.522258 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.526191 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"eeb3782c-0a88-4d17-b481-e2aef7ea386a","Type":"ContainerStarted","Data":"3df4a2193c9907b10329446f679b3d045c8e0838be144262dd416a7125f8a215"} Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.528286 4948 generic.go:334] "Generic (PLEG): container finished" podID="977992e9-30ca-4420-98ae-92f91545cd10" containerID="2fa62b0646099a9e4d67d43a31c1228c43fee77b21220b1456b925205d581d9a" exitCode=0 Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.528345 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" event={"ID":"977992e9-30ca-4420-98ae-92f91545cd10","Type":"ContainerDied","Data":"2fa62b0646099a9e4d67d43a31c1228c43fee77b21220b1456b925205d581d9a"} Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.528365 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" event={"ID":"977992e9-30ca-4420-98ae-92f91545cd10","Type":"ContainerDied","Data":"a0508d6a1f4b7f39846aac7f90bdfe83ad8793d4225f8bb5f2eaca0710969807"} Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.528436 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.532014 4948 generic.go:334] "Generic (PLEG): container finished" podID="33a06f41-ea8b-4138-8048-5425cd04f015" containerID="664c1ad747d8bd0a828811f06f4556fb8972cb9b47fe4b7d90fa3826c1e33b4a" exitCode=0 Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.532073 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" event={"ID":"33a06f41-ea8b-4138-8048-5425cd04f015","Type":"ContainerDied","Data":"664c1ad747d8bd0a828811f06f4556fb8972cb9b47fe4b7d90fa3826c1e33b4a"} Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.532099 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" event={"ID":"33a06f41-ea8b-4138-8048-5425cd04f015","Type":"ContainerStarted","Data":"1f7e5587a47afc9ffa01b7b40dc0afdae69e9922140a4a9e4162b6a2a7417196"} Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.533699 4948 generic.go:334] "Generic (PLEG): container finished" podID="8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be" containerID="c624ced0b3cd9432f64567b57ff22f472c7a407536a9a515e94a3ab95cf57662" exitCode=0 Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.534770 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" event={"ID":"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be","Type":"ContainerDied","Data":"c624ced0b3cd9432f64567b57ff22f472c7a407536a9a515e94a3ab95cf57662"} Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.534791 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" event={"ID":"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be","Type":"ContainerStarted","Data":"a0b4a492fe4751057387e9dea455a523858f53f3d884620fffd805ea3be59755"} Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.553252 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-gcfn8" podStartSLOduration=2.553236133 podStartE2EDuration="2.553236133s" podCreationTimestamp="2026-02-20 08:21:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:21:22.548639479 +0000 UTC m=+931.523134299" watchObservedRunningTime="2026-02-20 08:21:22.553236133 +0000 UTC m=+931.527730953" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.608455 4948 scope.go:117] "RemoveContainer" containerID="5a5bfff9ba7b803c56d99738536efa4fb27865dfba1fd58a67dc23b1d4bef9ed" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.615914 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/977992e9-30ca-4420-98ae-92f91545cd10-config\") pod \"977992e9-30ca-4420-98ae-92f91545cd10\" (UID: \"977992e9-30ca-4420-98ae-92f91545cd10\") " Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.616799 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pds7s\" (UniqueName: \"kubernetes.io/projected/977992e9-30ca-4420-98ae-92f91545cd10-kube-api-access-pds7s\") pod \"977992e9-30ca-4420-98ae-92f91545cd10\" (UID: \"977992e9-30ca-4420-98ae-92f91545cd10\") " Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.616839 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/977992e9-30ca-4420-98ae-92f91545cd10-dns-svc\") pod \"977992e9-30ca-4420-98ae-92f91545cd10\" (UID: \"977992e9-30ca-4420-98ae-92f91545cd10\") " Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.623757 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/977992e9-30ca-4420-98ae-92f91545cd10-kube-api-access-pds7s" (OuterVolumeSpecName: "kube-api-access-pds7s") pod "977992e9-30ca-4420-98ae-92f91545cd10" (UID: "977992e9-30ca-4420-98ae-92f91545cd10"). InnerVolumeSpecName "kube-api-access-pds7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.672508 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-mkkzf"] Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.675393 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/977992e9-30ca-4420-98ae-92f91545cd10-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "977992e9-30ca-4420-98ae-92f91545cd10" (UID: "977992e9-30ca-4420-98ae-92f91545cd10"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.684064 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/977992e9-30ca-4420-98ae-92f91545cd10-config" (OuterVolumeSpecName: "config") pod "977992e9-30ca-4420-98ae-92f91545cd10" (UID: "977992e9-30ca-4420-98ae-92f91545cd10"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.692460 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-mkkzf"] Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.693883 4948 scope.go:117] "RemoveContainer" containerID="2fa62b0646099a9e4d67d43a31c1228c43fee77b21220b1456b925205d581d9a" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.719952 4948 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/977992e9-30ca-4420-98ae-92f91545cd10-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.720004 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/977992e9-30ca-4420-98ae-92f91545cd10-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.720014 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pds7s\" (UniqueName: \"kubernetes.io/projected/977992e9-30ca-4420-98ae-92f91545cd10-kube-api-access-pds7s\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.724694 4948 scope.go:117] "RemoveContainer" containerID="05ccd3525150ab098e3561398cfda7a7a04a42e4d12ac43eacc78bfa91ca9099" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.755945 4948 scope.go:117] "RemoveContainer" containerID="2fa62b0646099a9e4d67d43a31c1228c43fee77b21220b1456b925205d581d9a" Feb 20 08:21:22 crc kubenswrapper[4948]: E0220 08:21:22.756659 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fa62b0646099a9e4d67d43a31c1228c43fee77b21220b1456b925205d581d9a\": container with ID starting with 2fa62b0646099a9e4d67d43a31c1228c43fee77b21220b1456b925205d581d9a not found: ID does not exist" containerID="2fa62b0646099a9e4d67d43a31c1228c43fee77b21220b1456b925205d581d9a" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.756685 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fa62b0646099a9e4d67d43a31c1228c43fee77b21220b1456b925205d581d9a"} err="failed to get container status \"2fa62b0646099a9e4d67d43a31c1228c43fee77b21220b1456b925205d581d9a\": rpc error: code = NotFound desc = could not find container \"2fa62b0646099a9e4d67d43a31c1228c43fee77b21220b1456b925205d581d9a\": container with ID starting with 2fa62b0646099a9e4d67d43a31c1228c43fee77b21220b1456b925205d581d9a not found: ID does not exist" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.756706 4948 scope.go:117] "RemoveContainer" containerID="05ccd3525150ab098e3561398cfda7a7a04a42e4d12ac43eacc78bfa91ca9099" Feb 20 08:21:22 crc kubenswrapper[4948]: E0220 08:21:22.757008 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05ccd3525150ab098e3561398cfda7a7a04a42e4d12ac43eacc78bfa91ca9099\": container with ID starting with 05ccd3525150ab098e3561398cfda7a7a04a42e4d12ac43eacc78bfa91ca9099 not found: ID does not exist" containerID="05ccd3525150ab098e3561398cfda7a7a04a42e4d12ac43eacc78bfa91ca9099" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.757026 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05ccd3525150ab098e3561398cfda7a7a04a42e4d12ac43eacc78bfa91ca9099"} err="failed to get container status \"05ccd3525150ab098e3561398cfda7a7a04a42e4d12ac43eacc78bfa91ca9099\": rpc error: code = NotFound desc = could not find container \"05ccd3525150ab098e3561398cfda7a7a04a42e4d12ac43eacc78bfa91ca9099\": container with ID starting with 05ccd3525150ab098e3561398cfda7a7a04a42e4d12ac43eacc78bfa91ca9099 not found: ID does not exist" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.921241 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zt6gg"] Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.927468 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zt6gg"] Feb 20 08:21:23 crc kubenswrapper[4948]: I0220 08:21:23.546058 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" event={"ID":"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be","Type":"ContainerStarted","Data":"ae779bfaf190f36bd8dce0329a6a4c0ac105d70cb79287b3be92631d581020d5"} Feb 20 08:21:23 crc kubenswrapper[4948]: I0220 08:21:23.546396 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:23 crc kubenswrapper[4948]: I0220 08:21:23.550231 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" event={"ID":"33a06f41-ea8b-4138-8048-5425cd04f015","Type":"ContainerStarted","Data":"489a052e4c0fb8182ef93444ec3a663e93256a532f88c70856200cb8dc8b254f"} Feb 20 08:21:23 crc kubenswrapper[4948]: I0220 08:21:23.569174 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" podStartSLOduration=3.569156304 podStartE2EDuration="3.569156304s" podCreationTimestamp="2026-02-20 08:21:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:21:23.566068538 +0000 UTC m=+932.540563358" watchObservedRunningTime="2026-02-20 08:21:23.569156304 +0000 UTC m=+932.543651134" Feb 20 08:21:23 crc kubenswrapper[4948]: I0220 08:21:23.596041 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" podStartSLOduration=3.5960244489999997 podStartE2EDuration="3.596024449s" podCreationTimestamp="2026-02-20 08:21:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:21:23.595136177 +0000 UTC m=+932.569631017" watchObservedRunningTime="2026-02-20 08:21:23.596024449 +0000 UTC m=+932.570519269" Feb 20 08:21:23 crc kubenswrapper[4948]: I0220 08:21:23.735682 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5686ae39-4e44-4303-9b44-ad4b9ceae1ba" path="/var/lib/kubelet/pods/5686ae39-4e44-4303-9b44-ad4b9ceae1ba/volumes" Feb 20 08:21:23 crc kubenswrapper[4948]: I0220 08:21:23.736385 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="977992e9-30ca-4420-98ae-92f91545cd10" path="/var/lib/kubelet/pods/977992e9-30ca-4420-98ae-92f91545cd10/volumes" Feb 20 08:21:24 crc kubenswrapper[4948]: I0220 08:21:24.557819 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"eeb3782c-0a88-4d17-b481-e2aef7ea386a","Type":"ContainerStarted","Data":"de9a37bb635341d457c0f9567c4af94f2752b0f6fb2a1c5b47992089cf013a01"} Feb 20 08:21:24 crc kubenswrapper[4948]: I0220 08:21:24.558095 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"eeb3782c-0a88-4d17-b481-e2aef7ea386a","Type":"ContainerStarted","Data":"ad5f6f0c61518eb7ded2c63c3aa9c6b2d5905b20178bc434822447b5304588eb"} Feb 20 08:21:24 crc kubenswrapper[4948]: I0220 08:21:24.558318 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:24 crc kubenswrapper[4948]: I0220 08:21:24.589662 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.128991371 podStartE2EDuration="4.589644578s" podCreationTimestamp="2026-02-20 08:21:20 +0000 UTC" firstStartedPulling="2026-02-20 08:21:21.87829986 +0000 UTC m=+930.852794680" lastFinishedPulling="2026-02-20 08:21:23.338953067 +0000 UTC m=+932.313447887" observedRunningTime="2026-02-20 08:21:24.582508822 +0000 UTC m=+933.557003642" watchObservedRunningTime="2026-02-20 08:21:24.589644578 +0000 UTC m=+933.564139398" Feb 20 08:21:25 crc kubenswrapper[4948]: I0220 08:21:25.519601 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Feb 20 08:21:25 crc kubenswrapper[4948]: I0220 08:21:25.519667 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Feb 20 08:21:25 crc kubenswrapper[4948]: I0220 08:21:25.567486 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Feb 20 08:21:25 crc kubenswrapper[4948]: I0220 08:21:25.634923 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Feb 20 08:21:25 crc kubenswrapper[4948]: I0220 08:21:25.720462 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Feb 20 08:21:25 crc kubenswrapper[4948]: I0220 08:21:25.853723 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rt8tp" Feb 20 08:21:25 crc kubenswrapper[4948]: I0220 08:21:25.854156 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rt8tp" Feb 20 08:21:25 crc kubenswrapper[4948]: I0220 08:21:25.920903 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rt8tp" Feb 20 08:21:26 crc kubenswrapper[4948]: I0220 08:21:26.617742 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rt8tp" Feb 20 08:21:26 crc kubenswrapper[4948]: I0220 08:21:26.683299 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rt8tp"] Feb 20 08:21:26 crc kubenswrapper[4948]: I0220 08:21:26.898074 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Feb 20 08:21:26 crc kubenswrapper[4948]: I0220 08:21:26.898407 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.106189 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-eb36-account-create-update-554g5"] Feb 20 08:21:28 crc kubenswrapper[4948]: E0220 08:21:28.106721 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="977992e9-30ca-4420-98ae-92f91545cd10" containerName="init" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.106733 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="977992e9-30ca-4420-98ae-92f91545cd10" containerName="init" Feb 20 08:21:28 crc kubenswrapper[4948]: E0220 08:21:28.106752 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5686ae39-4e44-4303-9b44-ad4b9ceae1ba" containerName="init" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.106758 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="5686ae39-4e44-4303-9b44-ad4b9ceae1ba" containerName="init" Feb 20 08:21:28 crc kubenswrapper[4948]: E0220 08:21:28.106773 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="977992e9-30ca-4420-98ae-92f91545cd10" containerName="dnsmasq-dns" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.106778 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="977992e9-30ca-4420-98ae-92f91545cd10" containerName="dnsmasq-dns" Feb 20 08:21:28 crc kubenswrapper[4948]: E0220 08:21:28.106798 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5686ae39-4e44-4303-9b44-ad4b9ceae1ba" containerName="dnsmasq-dns" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.106803 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="5686ae39-4e44-4303-9b44-ad4b9ceae1ba" containerName="dnsmasq-dns" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.106955 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="5686ae39-4e44-4303-9b44-ad4b9ceae1ba" containerName="dnsmasq-dns" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.106987 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="977992e9-30ca-4420-98ae-92f91545cd10" containerName="dnsmasq-dns" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.107422 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-eb36-account-create-update-554g5" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.109040 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.115191 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-eb36-account-create-update-554g5"] Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.174666 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-gm6g6"] Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.175728 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gm6g6" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.184786 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-gm6g6"] Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.253653 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5bf7a21-af5d-4334-bc98-bd0f9eb9035a-operator-scripts\") pod \"keystone-eb36-account-create-update-554g5\" (UID: \"b5bf7a21-af5d-4334-bc98-bd0f9eb9035a\") " pod="openstack/keystone-eb36-account-create-update-554g5" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.253729 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4wx7\" (UniqueName: \"kubernetes.io/projected/b5bf7a21-af5d-4334-bc98-bd0f9eb9035a-kube-api-access-j4wx7\") pod \"keystone-eb36-account-create-update-554g5\" (UID: \"b5bf7a21-af5d-4334-bc98-bd0f9eb9035a\") " pod="openstack/keystone-eb36-account-create-update-554g5" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.278208 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-bv65d"] Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.279168 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-bv65d" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.289501 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-bv65d"] Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.302183 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.355659 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89k8g\" (UniqueName: \"kubernetes.io/projected/52238644-5964-427e-8b77-9ca228d031db-kube-api-access-89k8g\") pod \"keystone-db-create-gm6g6\" (UID: \"52238644-5964-427e-8b77-9ca228d031db\") " pod="openstack/keystone-db-create-gm6g6" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.355793 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52238644-5964-427e-8b77-9ca228d031db-operator-scripts\") pod \"keystone-db-create-gm6g6\" (UID: \"52238644-5964-427e-8b77-9ca228d031db\") " pod="openstack/keystone-db-create-gm6g6" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.355984 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5bf7a21-af5d-4334-bc98-bd0f9eb9035a-operator-scripts\") pod \"keystone-eb36-account-create-update-554g5\" (UID: \"b5bf7a21-af5d-4334-bc98-bd0f9eb9035a\") " pod="openstack/keystone-eb36-account-create-update-554g5" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.356094 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4wx7\" (UniqueName: \"kubernetes.io/projected/b5bf7a21-af5d-4334-bc98-bd0f9eb9035a-kube-api-access-j4wx7\") pod \"keystone-eb36-account-create-update-554g5\" (UID: \"b5bf7a21-af5d-4334-bc98-bd0f9eb9035a\") " pod="openstack/keystone-eb36-account-create-update-554g5" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.357281 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5bf7a21-af5d-4334-bc98-bd0f9eb9035a-operator-scripts\") pod \"keystone-eb36-account-create-update-554g5\" (UID: \"b5bf7a21-af5d-4334-bc98-bd0f9eb9035a\") " pod="openstack/keystone-eb36-account-create-update-554g5" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.383395 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6d58-account-create-update-bbx44"] Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.384609 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4wx7\" (UniqueName: \"kubernetes.io/projected/b5bf7a21-af5d-4334-bc98-bd0f9eb9035a-kube-api-access-j4wx7\") pod \"keystone-eb36-account-create-update-554g5\" (UID: \"b5bf7a21-af5d-4334-bc98-bd0f9eb9035a\") " pod="openstack/keystone-eb36-account-create-update-554g5" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.384808 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6d58-account-create-update-bbx44" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.393184 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.410424 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6d58-account-create-update-bbx44"] Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.425042 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-eb36-account-create-update-554g5" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.457720 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89k8g\" (UniqueName: \"kubernetes.io/projected/52238644-5964-427e-8b77-9ca228d031db-kube-api-access-89k8g\") pod \"keystone-db-create-gm6g6\" (UID: \"52238644-5964-427e-8b77-9ca228d031db\") " pod="openstack/keystone-db-create-gm6g6" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.457793 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52238644-5964-427e-8b77-9ca228d031db-operator-scripts\") pod \"keystone-db-create-gm6g6\" (UID: \"52238644-5964-427e-8b77-9ca228d031db\") " pod="openstack/keystone-db-create-gm6g6" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.457920 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f917fa7f-51f9-4c8f-bee2-11529a6d44a6-operator-scripts\") pod \"placement-db-create-bv65d\" (UID: \"f917fa7f-51f9-4c8f-bee2-11529a6d44a6\") " pod="openstack/placement-db-create-bv65d" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.457983 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnscr\" (UniqueName: \"kubernetes.io/projected/f917fa7f-51f9-4c8f-bee2-11529a6d44a6-kube-api-access-wnscr\") pod \"placement-db-create-bv65d\" (UID: \"f917fa7f-51f9-4c8f-bee2-11529a6d44a6\") " pod="openstack/placement-db-create-bv65d" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.458818 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52238644-5964-427e-8b77-9ca228d031db-operator-scripts\") pod \"keystone-db-create-gm6g6\" (UID: \"52238644-5964-427e-8b77-9ca228d031db\") " pod="openstack/keystone-db-create-gm6g6" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.473910 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89k8g\" (UniqueName: \"kubernetes.io/projected/52238644-5964-427e-8b77-9ca228d031db-kube-api-access-89k8g\") pod \"keystone-db-create-gm6g6\" (UID: \"52238644-5964-427e-8b77-9ca228d031db\") " pod="openstack/keystone-db-create-gm6g6" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.513907 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gm6g6" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.560108 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnscr\" (UniqueName: \"kubernetes.io/projected/f917fa7f-51f9-4c8f-bee2-11529a6d44a6-kube-api-access-wnscr\") pod \"placement-db-create-bv65d\" (UID: \"f917fa7f-51f9-4c8f-bee2-11529a6d44a6\") " pod="openstack/placement-db-create-bv65d" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.560442 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kqvl\" (UniqueName: \"kubernetes.io/projected/e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb-kube-api-access-2kqvl\") pod \"placement-6d58-account-create-update-bbx44\" (UID: \"e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb\") " pod="openstack/placement-6d58-account-create-update-bbx44" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.560579 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb-operator-scripts\") pod \"placement-6d58-account-create-update-bbx44\" (UID: \"e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb\") " pod="openstack/placement-6d58-account-create-update-bbx44" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.561148 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f917fa7f-51f9-4c8f-bee2-11529a6d44a6-operator-scripts\") pod \"placement-db-create-bv65d\" (UID: \"f917fa7f-51f9-4c8f-bee2-11529a6d44a6\") " pod="openstack/placement-db-create-bv65d" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.561727 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f917fa7f-51f9-4c8f-bee2-11529a6d44a6-operator-scripts\") pod \"placement-db-create-bv65d\" (UID: \"f917fa7f-51f9-4c8f-bee2-11529a6d44a6\") " pod="openstack/placement-db-create-bv65d" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.595819 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnscr\" (UniqueName: \"kubernetes.io/projected/f917fa7f-51f9-4c8f-bee2-11529a6d44a6-kube-api-access-wnscr\") pod \"placement-db-create-bv65d\" (UID: \"f917fa7f-51f9-4c8f-bee2-11529a6d44a6\") " pod="openstack/placement-db-create-bv65d" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.598034 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rt8tp" podUID="93f24735-d27d-4867-b9f7-cce47e7f376f" containerName="registry-server" containerID="cri-o://fbf3b4c07c31284ef058431c15f6fb31624a27a446740ecaf65fcf08b863a6e3" gracePeriod=2 Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.664103 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb-operator-scripts\") pod \"placement-6d58-account-create-update-bbx44\" (UID: \"e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb\") " pod="openstack/placement-6d58-account-create-update-bbx44" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.665652 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb-operator-scripts\") pod \"placement-6d58-account-create-update-bbx44\" (UID: \"e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb\") " pod="openstack/placement-6d58-account-create-update-bbx44" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.665846 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kqvl\" (UniqueName: \"kubernetes.io/projected/e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb-kube-api-access-2kqvl\") pod \"placement-6d58-account-create-update-bbx44\" (UID: \"e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb\") " pod="openstack/placement-6d58-account-create-update-bbx44" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.687287 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kqvl\" (UniqueName: \"kubernetes.io/projected/e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb-kube-api-access-2kqvl\") pod \"placement-6d58-account-create-update-bbx44\" (UID: \"e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb\") " pod="openstack/placement-6d58-account-create-update-bbx44" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.705247 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.722027 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6d58-account-create-update-bbx44" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.894340 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-bv65d" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.905789 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-eb36-account-create-update-554g5"] Feb 20 08:21:28 crc kubenswrapper[4948]: W0220 08:21:28.912934 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5bf7a21_af5d_4334_bc98_bd0f9eb9035a.slice/crio-25d97fb00e18b6ce84c602bbc6e333780c0537c4cab9cb2b3ce67336516b898c WatchSource:0}: Error finding container 25d97fb00e18b6ce84c602bbc6e333780c0537c4cab9cb2b3ce67336516b898c: Status 404 returned error can't find the container with id 25d97fb00e18b6ce84c602bbc6e333780c0537c4cab9cb2b3ce67336516b898c Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.038591 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-gm6g6"] Feb 20 08:21:29 crc kubenswrapper[4948]: W0220 08:21:29.067824 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52238644_5964_427e_8b77_9ca228d031db.slice/crio-79852731770fecb861112fde754b01ca4c49d355a45bb283f9e41b055faf6747 WatchSource:0}: Error finding container 79852731770fecb861112fde754b01ca4c49d355a45bb283f9e41b055faf6747: Status 404 returned error can't find the container with id 79852731770fecb861112fde754b01ca4c49d355a45bb283f9e41b055faf6747 Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.173853 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6d58-account-create-update-bbx44"] Feb 20 08:21:29 crc kubenswrapper[4948]: W0220 08:21:29.181928 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1700d74_b484_4c6e_8fd3_d9dab8f4b7eb.slice/crio-fe0235453e94ed6b67ebdad926c4872aff88ca5acd235d564a2517d91ee0dafd WatchSource:0}: Error finding container fe0235453e94ed6b67ebdad926c4872aff88ca5acd235d564a2517d91ee0dafd: Status 404 returned error can't find the container with id fe0235453e94ed6b67ebdad926c4872aff88ca5acd235d564a2517d91ee0dafd Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.373185 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-bv65d"] Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.387251 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 20 08:21:29 crc kubenswrapper[4948]: W0220 08:21:29.396025 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf917fa7f_51f9_4c8f_bee2_11529a6d44a6.slice/crio-7f25c2e14bdc30bc83d90e8cf3dae79ddb8cda1ee4900ffa7eb5001c2eb0fd98 WatchSource:0}: Error finding container 7f25c2e14bdc30bc83d90e8cf3dae79ddb8cda1ee4900ffa7eb5001c2eb0fd98: Status 404 returned error can't find the container with id 7f25c2e14bdc30bc83d90e8cf3dae79ddb8cda1ee4900ffa7eb5001c2eb0fd98 Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.463706 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-mvb6f"] Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.463916 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" podUID="8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be" containerName="dnsmasq-dns" containerID="cri-o://ae779bfaf190f36bd8dce0329a6a4c0ac105d70cb79287b3be92631d581020d5" gracePeriod=10 Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.465427 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.514041 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-tkqzb"] Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.515708 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.532520 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-tkqzb"] Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.610332 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gm6g6" event={"ID":"52238644-5964-427e-8b77-9ca228d031db","Type":"ContainerStarted","Data":"79852731770fecb861112fde754b01ca4c49d355a45bb283f9e41b055faf6747"} Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.611161 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-eb36-account-create-update-554g5" event={"ID":"b5bf7a21-af5d-4334-bc98-bd0f9eb9035a","Type":"ContainerStarted","Data":"25d97fb00e18b6ce84c602bbc6e333780c0537c4cab9cb2b3ce67336516b898c"} Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.612324 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6d58-account-create-update-bbx44" event={"ID":"e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb","Type":"ContainerStarted","Data":"fe0235453e94ed6b67ebdad926c4872aff88ca5acd235d564a2517d91ee0dafd"} Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.613025 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-bv65d" event={"ID":"f917fa7f-51f9-4c8f-bee2-11529a6d44a6","Type":"ContainerStarted","Data":"7f25c2e14bdc30bc83d90e8cf3dae79ddb8cda1ee4900ffa7eb5001c2eb0fd98"} Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.615492 4948 generic.go:334] "Generic (PLEG): container finished" podID="93f24735-d27d-4867-b9f7-cce47e7f376f" containerID="fbf3b4c07c31284ef058431c15f6fb31624a27a446740ecaf65fcf08b863a6e3" exitCode=0 Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.615566 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rt8tp" event={"ID":"93f24735-d27d-4867-b9f7-cce47e7f376f","Type":"ContainerDied","Data":"fbf3b4c07c31284ef058431c15f6fb31624a27a446740ecaf65fcf08b863a6e3"} Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.632386 4948 generic.go:334] "Generic (PLEG): container finished" podID="8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be" containerID="ae779bfaf190f36bd8dce0329a6a4c0ac105d70cb79287b3be92631d581020d5" exitCode=0 Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.633247 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" event={"ID":"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be","Type":"ContainerDied","Data":"ae779bfaf190f36bd8dce0329a6a4c0ac105d70cb79287b3be92631d581020d5"} Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.686792 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wc2m\" (UniqueName: \"kubernetes.io/projected/da816a64-160f-4a5c-b945-b8fa467208d6-kube-api-access-4wc2m\") pod \"dnsmasq-dns-698758b865-tkqzb\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.687177 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-config\") pod \"dnsmasq-dns-698758b865-tkqzb\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.687208 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-dns-svc\") pod \"dnsmasq-dns-698758b865-tkqzb\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.687235 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-tkqzb\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.687336 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-tkqzb\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.788225 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-config\") pod \"dnsmasq-dns-698758b865-tkqzb\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.788650 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-dns-svc\") pod \"dnsmasq-dns-698758b865-tkqzb\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.788676 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-tkqzb\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.789266 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-config\") pod \"dnsmasq-dns-698758b865-tkqzb\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.789376 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-dns-svc\") pod \"dnsmasq-dns-698758b865-tkqzb\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.789647 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-tkqzb\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.790079 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-tkqzb\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.790176 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wc2m\" (UniqueName: \"kubernetes.io/projected/da816a64-160f-4a5c-b945-b8fa467208d6-kube-api-access-4wc2m\") pod \"dnsmasq-dns-698758b865-tkqzb\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.790617 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-tkqzb\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.808625 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wc2m\" (UniqueName: \"kubernetes.io/projected/da816a64-160f-4a5c-b945-b8fa467208d6-kube-api-access-4wc2m\") pod \"dnsmasq-dns-698758b865-tkqzb\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.847300 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.274959 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-tkqzb"] Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.427412 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rt8tp" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.589560 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Feb 20 08:21:30 crc kubenswrapper[4948]: E0220 08:21:30.596366 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93f24735-d27d-4867-b9f7-cce47e7f376f" containerName="extract-utilities" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.596399 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="93f24735-d27d-4867-b9f7-cce47e7f376f" containerName="extract-utilities" Feb 20 08:21:30 crc kubenswrapper[4948]: E0220 08:21:30.596413 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93f24735-d27d-4867-b9f7-cce47e7f376f" containerName="registry-server" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.596419 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="93f24735-d27d-4867-b9f7-cce47e7f376f" containerName="registry-server" Feb 20 08:21:30 crc kubenswrapper[4948]: E0220 08:21:30.596429 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93f24735-d27d-4867-b9f7-cce47e7f376f" containerName="extract-content" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.596435 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="93f24735-d27d-4867-b9f7-cce47e7f376f" containerName="extract-content" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.596579 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="93f24735-d27d-4867-b9f7-cce47e7f376f" containerName="registry-server" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.601196 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.607266 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.607391 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.607419 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-8lns4" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.607604 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.608909 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93f24735-d27d-4867-b9f7-cce47e7f376f-catalog-content\") pod \"93f24735-d27d-4867-b9f7-cce47e7f376f\" (UID: \"93f24735-d27d-4867-b9f7-cce47e7f376f\") " Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.608965 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2cfc\" (UniqueName: \"kubernetes.io/projected/93f24735-d27d-4867-b9f7-cce47e7f376f-kube-api-access-q2cfc\") pod \"93f24735-d27d-4867-b9f7-cce47e7f376f\" (UID: \"93f24735-d27d-4867-b9f7-cce47e7f376f\") " Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.609081 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93f24735-d27d-4867-b9f7-cce47e7f376f-utilities\") pod \"93f24735-d27d-4867-b9f7-cce47e7f376f\" (UID: \"93f24735-d27d-4867-b9f7-cce47e7f376f\") " Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.610148 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93f24735-d27d-4867-b9f7-cce47e7f376f-utilities" (OuterVolumeSpecName: "utilities") pod "93f24735-d27d-4867-b9f7-cce47e7f376f" (UID: "93f24735-d27d-4867-b9f7-cce47e7f376f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.612607 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.619171 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93f24735-d27d-4867-b9f7-cce47e7f376f-kube-api-access-q2cfc" (OuterVolumeSpecName: "kube-api-access-q2cfc") pod "93f24735-d27d-4867-b9f7-cce47e7f376f" (UID: "93f24735-d27d-4867-b9f7-cce47e7f376f"). InnerVolumeSpecName "kube-api-access-q2cfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.648811 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-eb36-account-create-update-554g5" event={"ID":"b5bf7a21-af5d-4334-bc98-bd0f9eb9035a","Type":"ContainerStarted","Data":"f4eb6ffc3a340bf3fb020244f7c41526c57d80da4f478c0885a3916041b2da30"} Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.660057 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6d58-account-create-update-bbx44" event={"ID":"e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb","Type":"ContainerStarted","Data":"043fb8ebd535abc3ac64988e423e5b76f657b01785d57b024ae113afd44b4bf7"} Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.670809 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-tkqzb" event={"ID":"da816a64-160f-4a5c-b945-b8fa467208d6","Type":"ContainerStarted","Data":"32b2c84d98711b304188499ff3c33cd5bf7af6d44c7a701d021f0522c169184b"} Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.680318 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-bv65d" event={"ID":"f917fa7f-51f9-4c8f-bee2-11529a6d44a6","Type":"ContainerStarted","Data":"279cb01f22f256d8185916f133561b9dbb69d1fb0ef3e531faed5ee14470b9d2"} Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.682624 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rt8tp" event={"ID":"93f24735-d27d-4867-b9f7-cce47e7f376f","Type":"ContainerDied","Data":"104e51617975f8f0d6bf864e106e2486e0a18298cc9a24f078002083a6d423b2"} Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.682668 4948 scope.go:117] "RemoveContainer" containerID="fbf3b4c07c31284ef058431c15f6fb31624a27a446740ecaf65fcf08b863a6e3" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.682755 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rt8tp" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.687950 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gm6g6" event={"ID":"52238644-5964-427e-8b77-9ca228d031db","Type":"ContainerStarted","Data":"eb75184d7c11178b70689bbef778a20eca59576d9cf6e2826963a0598aa904a4"} Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.701068 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-eb36-account-create-update-554g5" podStartSLOduration=2.701052507 podStartE2EDuration="2.701052507s" podCreationTimestamp="2026-02-20 08:21:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:21:30.699405706 +0000 UTC m=+939.673900526" watchObservedRunningTime="2026-02-20 08:21:30.701052507 +0000 UTC m=+939.675547327" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.709339 4948 scope.go:117] "RemoveContainer" containerID="7f925a093fb8a5b68097ec9940401bb5483d01b61401a6a70fda1f29efe6615e" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.710229 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.710283 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4d60def-5dc9-460b-92ed-e3d0157982ec-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.710314 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/e4d60def-5dc9-460b-92ed-e3d0157982ec-cache\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.710362 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.710384 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvm4h\" (UniqueName: \"kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-kube-api-access-vvm4h\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.710403 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/e4d60def-5dc9-460b-92ed-e3d0157982ec-lock\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.710439 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2cfc\" (UniqueName: \"kubernetes.io/projected/93f24735-d27d-4867-b9f7-cce47e7f376f-kube-api-access-q2cfc\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.710449 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93f24735-d27d-4867-b9f7-cce47e7f376f-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.735463 4948 scope.go:117] "RemoveContainer" containerID="b689fa43e3b5c02fa357d00443dc58a7d6fa6d71f6692b435028801366e8a5e6" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.738413 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-6d58-account-create-update-bbx44" podStartSLOduration=2.738398541 podStartE2EDuration="2.738398541s" podCreationTimestamp="2026-02-20 08:21:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:21:30.733635633 +0000 UTC m=+939.708130453" watchObservedRunningTime="2026-02-20 08:21:30.738398541 +0000 UTC m=+939.712893361" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.765879 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-bv65d" podStartSLOduration=2.76585781 podStartE2EDuration="2.76585781s" podCreationTimestamp="2026-02-20 08:21:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:21:30.755720539 +0000 UTC m=+939.730215359" watchObservedRunningTime="2026-02-20 08:21:30.76585781 +0000 UTC m=+939.740352630" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.793904 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-gm6g6" podStartSLOduration=2.793887644 podStartE2EDuration="2.793887644s" podCreationTimestamp="2026-02-20 08:21:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:21:30.780591955 +0000 UTC m=+939.755086775" watchObservedRunningTime="2026-02-20 08:21:30.793887644 +0000 UTC m=+939.768382464" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.812524 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.812591 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvm4h\" (UniqueName: \"kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-kube-api-access-vvm4h\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.812621 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/e4d60def-5dc9-460b-92ed-e3d0157982ec-lock\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.812671 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.812771 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4d60def-5dc9-460b-92ed-e3d0157982ec-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.812813 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/e4d60def-5dc9-460b-92ed-e3d0157982ec-cache\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.813697 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: E0220 08:21:30.816528 4948 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 20 08:21:30 crc kubenswrapper[4948]: E0220 08:21:30.816552 4948 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 20 08:21:30 crc kubenswrapper[4948]: E0220 08:21:30.816600 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift podName:e4d60def-5dc9-460b-92ed-e3d0157982ec nodeName:}" failed. No retries permitted until 2026-02-20 08:21:31.316584146 +0000 UTC m=+940.291078966 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift") pod "swift-storage-0" (UID: "e4d60def-5dc9-460b-92ed-e3d0157982ec") : configmap "swift-ring-files" not found Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.816828 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/e4d60def-5dc9-460b-92ed-e3d0157982ec-lock\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.826121 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/e4d60def-5dc9-460b-92ed-e3d0157982ec-cache\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.826774 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4d60def-5dc9-460b-92ed-e3d0157982ec-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.832184 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvm4h\" (UniqueName: \"kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-kube-api-access-vvm4h\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.839922 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.854347 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93f24735-d27d-4867-b9f7-cce47e7f376f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "93f24735-d27d-4867-b9f7-cce47e7f376f" (UID: "93f24735-d27d-4867-b9f7-cce47e7f376f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.913904 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93f24735-d27d-4867-b9f7-cce47e7f376f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.043526 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-f44f4"] Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.044472 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.046043 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.046344 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.046534 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.088113 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-f44f4"] Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.101483 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-f44f4"] Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.108434 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-dllls"] Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.109708 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.117127 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-combined-ca-bundle\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.117202 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-swiftconf\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.117270 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/adf79be9-a443-4d51-b854-2067624fc60f-scripts\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.117381 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px2hh\" (UniqueName: \"kubernetes.io/projected/adf79be9-a443-4d51-b854-2067624fc60f-kube-api-access-px2hh\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.117518 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/adf79be9-a443-4d51-b854-2067624fc60f-etc-swift\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.117598 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/adf79be9-a443-4d51-b854-2067624fc60f-ring-data-devices\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.117660 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-dispersionconf\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.129083 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-dllls"] Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.137656 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:31 crc kubenswrapper[4948]: E0220 08:21:31.150645 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-px2hh ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/swift-ring-rebalance-f44f4" podUID="adf79be9-a443-4d51-b854-2067624fc60f" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.162378 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rt8tp"] Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.168656 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rt8tp"] Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.219566 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-dns-svc\") pod \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\" (UID: \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\") " Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.219717 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-ovsdbserver-nb\") pod \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\" (UID: \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\") " Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.219738 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-config\") pod \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\" (UID: \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\") " Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.219772 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9cpkv\" (UniqueName: \"kubernetes.io/projected/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-kube-api-access-9cpkv\") pod \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\" (UID: \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\") " Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.220006 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-scripts\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.220036 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-swiftconf\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.220084 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-dispersionconf\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.220104 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/adf79be9-a443-4d51-b854-2067624fc60f-scripts\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.220120 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px2hh\" (UniqueName: \"kubernetes.io/projected/adf79be9-a443-4d51-b854-2067624fc60f-kube-api-access-px2hh\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.220143 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-ring-data-devices\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.220178 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/adf79be9-a443-4d51-b854-2067624fc60f-etc-swift\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.220199 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/adf79be9-a443-4d51-b854-2067624fc60f-ring-data-devices\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.220215 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-dispersionconf\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.220228 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-etc-swift\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.220256 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-combined-ca-bundle\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.220272 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89jgn\" (UniqueName: \"kubernetes.io/projected/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-kube-api-access-89jgn\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.220296 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-swiftconf\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.220331 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-combined-ca-bundle\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.221549 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/adf79be9-a443-4d51-b854-2067624fc60f-etc-swift\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.221747 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/adf79be9-a443-4d51-b854-2067624fc60f-ring-data-devices\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.222000 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/adf79be9-a443-4d51-b854-2067624fc60f-scripts\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.224097 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-dispersionconf\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.225314 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-swiftconf\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.226301 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-combined-ca-bundle\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.233758 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-kube-api-access-9cpkv" (OuterVolumeSpecName: "kube-api-access-9cpkv") pod "8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be" (UID: "8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be"). InnerVolumeSpecName "kube-api-access-9cpkv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.237485 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px2hh\" (UniqueName: \"kubernetes.io/projected/adf79be9-a443-4d51-b854-2067624fc60f-kube-api-access-px2hh\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.258033 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be" (UID: "8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:31 crc kubenswrapper[4948]: E0220 08:21:31.263238 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-config podName:8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be nodeName:}" failed. No retries permitted until 2026-02-20 08:21:31.763198628 +0000 UTC m=+940.737693448 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config" (UniqueName: "kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-config") pod "8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be" (UID: "8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be") : error deleting /var/lib/kubelet/pods/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be/volume-subpaths: remove /var/lib/kubelet/pods/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be/volume-subpaths: no such file or directory Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.263536 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be" (UID: "8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.322210 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-ring-data-devices\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.322256 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:31 crc kubenswrapper[4948]: E0220 08:21:31.322376 4948 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 20 08:21:31 crc kubenswrapper[4948]: E0220 08:21:31.322396 4948 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 20 08:21:31 crc kubenswrapper[4948]: E0220 08:21:31.322439 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift podName:e4d60def-5dc9-460b-92ed-e3d0157982ec nodeName:}" failed. No retries permitted until 2026-02-20 08:21:32.322424044 +0000 UTC m=+941.296918864 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift") pod "swift-storage-0" (UID: "e4d60def-5dc9-460b-92ed-e3d0157982ec") : configmap "swift-ring-files" not found Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.322795 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-etc-swift\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.322906 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-ring-data-devices\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.323056 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-etc-swift\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.323110 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-combined-ca-bundle\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.323129 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89jgn\" (UniqueName: \"kubernetes.io/projected/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-kube-api-access-89jgn\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.323691 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-swiftconf\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.323783 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-scripts\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.323845 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-dispersionconf\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.323988 4948 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.324000 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.324011 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9cpkv\" (UniqueName: \"kubernetes.io/projected/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-kube-api-access-9cpkv\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.324279 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-scripts\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.333410 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-dispersionconf\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.333549 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-swiftconf\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.333734 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-combined-ca-bundle\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.346101 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89jgn\" (UniqueName: \"kubernetes.io/projected/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-kube-api-access-89jgn\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.376184 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.463123 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.712166 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.712307 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" event={"ID":"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be","Type":"ContainerDied","Data":"a0b4a492fe4751057387e9dea455a523858f53f3d884620fffd805ea3be59755"} Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.712653 4948 scope.go:117] "RemoveContainer" containerID="ae779bfaf190f36bd8dce0329a6a4c0ac105d70cb79287b3be92631d581020d5" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.714015 4948 generic.go:334] "Generic (PLEG): container finished" podID="52238644-5964-427e-8b77-9ca228d031db" containerID="eb75184d7c11178b70689bbef778a20eca59576d9cf6e2826963a0598aa904a4" exitCode=0 Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.714155 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gm6g6" event={"ID":"52238644-5964-427e-8b77-9ca228d031db","Type":"ContainerDied","Data":"eb75184d7c11178b70689bbef778a20eca59576d9cf6e2826963a0598aa904a4"} Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.716626 4948 generic.go:334] "Generic (PLEG): container finished" podID="b5bf7a21-af5d-4334-bc98-bd0f9eb9035a" containerID="f4eb6ffc3a340bf3fb020244f7c41526c57d80da4f478c0885a3916041b2da30" exitCode=0 Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.716749 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-eb36-account-create-update-554g5" event={"ID":"b5bf7a21-af5d-4334-bc98-bd0f9eb9035a","Type":"ContainerDied","Data":"f4eb6ffc3a340bf3fb020244f7c41526c57d80da4f478c0885a3916041b2da30"} Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.718301 4948 generic.go:334] "Generic (PLEG): container finished" podID="e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb" containerID="043fb8ebd535abc3ac64988e423e5b76f657b01785d57b024ae113afd44b4bf7" exitCode=0 Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.718346 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6d58-account-create-update-bbx44" event={"ID":"e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb","Type":"ContainerDied","Data":"043fb8ebd535abc3ac64988e423e5b76f657b01785d57b024ae113afd44b4bf7"} Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.720119 4948 generic.go:334] "Generic (PLEG): container finished" podID="da816a64-160f-4a5c-b945-b8fa467208d6" containerID="48592a00a61a9ce774a176445c59d2e205743cbfe0a6d5b932c49a8ff31d312a" exitCode=0 Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.720197 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-tkqzb" event={"ID":"da816a64-160f-4a5c-b945-b8fa467208d6","Type":"ContainerDied","Data":"48592a00a61a9ce774a176445c59d2e205743cbfe0a6d5b932c49a8ff31d312a"} Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.721784 4948 generic.go:334] "Generic (PLEG): container finished" podID="f917fa7f-51f9-4c8f-bee2-11529a6d44a6" containerID="279cb01f22f256d8185916f133561b9dbb69d1fb0ef3e531faed5ee14470b9d2" exitCode=0 Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.721931 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.736403 4948 scope.go:117] "RemoveContainer" containerID="c624ced0b3cd9432f64567b57ff22f472c7a407536a9a515e94a3ab95cf57662" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.745178 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93f24735-d27d-4867-b9f7-cce47e7f376f" path="/var/lib/kubelet/pods/93f24735-d27d-4867-b9f7-cce47e7f376f/volumes" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.746839 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.829083 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-bv65d" event={"ID":"f917fa7f-51f9-4c8f-bee2-11529a6d44a6","Type":"ContainerDied","Data":"279cb01f22f256d8185916f133561b9dbb69d1fb0ef3e531faed5ee14470b9d2"} Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.837881 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/adf79be9-a443-4d51-b854-2067624fc60f-etc-swift\") pod \"adf79be9-a443-4d51-b854-2067624fc60f\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.837930 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/adf79be9-a443-4d51-b854-2067624fc60f-ring-data-devices\") pod \"adf79be9-a443-4d51-b854-2067624fc60f\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.838042 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-px2hh\" (UniqueName: \"kubernetes.io/projected/adf79be9-a443-4d51-b854-2067624fc60f-kube-api-access-px2hh\") pod \"adf79be9-a443-4d51-b854-2067624fc60f\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.838074 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-combined-ca-bundle\") pod \"adf79be9-a443-4d51-b854-2067624fc60f\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.838130 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-swiftconf\") pod \"adf79be9-a443-4d51-b854-2067624fc60f\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.838166 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/adf79be9-a443-4d51-b854-2067624fc60f-scripts\") pod \"adf79be9-a443-4d51-b854-2067624fc60f\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.838229 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-config\") pod \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\" (UID: \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\") " Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.838250 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-dispersionconf\") pod \"adf79be9-a443-4d51-b854-2067624fc60f\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.838949 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/adf79be9-a443-4d51-b854-2067624fc60f-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "adf79be9-a443-4d51-b854-2067624fc60f" (UID: "adf79be9-a443-4d51-b854-2067624fc60f"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.839256 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adf79be9-a443-4d51-b854-2067624fc60f-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "adf79be9-a443-4d51-b854-2067624fc60f" (UID: "adf79be9-a443-4d51-b854-2067624fc60f"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.839910 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-config" (OuterVolumeSpecName: "config") pod "8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be" (UID: "8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.840909 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/adf79be9-a443-4d51-b854-2067624fc60f-scripts" (OuterVolumeSpecName: "scripts") pod "adf79be9-a443-4d51-b854-2067624fc60f" (UID: "adf79be9-a443-4d51-b854-2067624fc60f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.842320 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "adf79be9-a443-4d51-b854-2067624fc60f" (UID: "adf79be9-a443-4d51-b854-2067624fc60f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.843062 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "adf79be9-a443-4d51-b854-2067624fc60f" (UID: "adf79be9-a443-4d51-b854-2067624fc60f"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.843945 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "adf79be9-a443-4d51-b854-2067624fc60f" (UID: "adf79be9-a443-4d51-b854-2067624fc60f"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.845088 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adf79be9-a443-4d51-b854-2067624fc60f-kube-api-access-px2hh" (OuterVolumeSpecName: "kube-api-access-px2hh") pod "adf79be9-a443-4d51-b854-2067624fc60f" (UID: "adf79be9-a443-4d51-b854-2067624fc60f"). InnerVolumeSpecName "kube-api-access-px2hh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.903877 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-dllls"] Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.940923 4948 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/adf79be9-a443-4d51-b854-2067624fc60f-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.940962 4948 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/adf79be9-a443-4d51-b854-2067624fc60f-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.940991 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-px2hh\" (UniqueName: \"kubernetes.io/projected/adf79be9-a443-4d51-b854-2067624fc60f-kube-api-access-px2hh\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.941003 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.941014 4948 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.941027 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/adf79be9-a443-4d51-b854-2067624fc60f-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.941038 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.941049 4948 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.050380 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-mvb6f"] Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.058408 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-mvb6f"] Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.319821 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-fs9rr"] Feb 20 08:21:32 crc kubenswrapper[4948]: E0220 08:21:32.320276 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be" containerName="init" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.320303 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be" containerName="init" Feb 20 08:21:32 crc kubenswrapper[4948]: E0220 08:21:32.320327 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be" containerName="dnsmasq-dns" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.320338 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be" containerName="dnsmasq-dns" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.320586 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be" containerName="dnsmasq-dns" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.321324 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-fs9rr" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.341550 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-fs9rr"] Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.349510 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:32 crc kubenswrapper[4948]: E0220 08:21:32.349783 4948 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 20 08:21:32 crc kubenswrapper[4948]: E0220 08:21:32.349816 4948 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 20 08:21:32 crc kubenswrapper[4948]: E0220 08:21:32.349887 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift podName:e4d60def-5dc9-460b-92ed-e3d0157982ec nodeName:}" failed. No retries permitted until 2026-02-20 08:21:34.349860869 +0000 UTC m=+943.324355719 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift") pod "swift-storage-0" (UID: "e4d60def-5dc9-460b-92ed-e3d0157982ec") : configmap "swift-ring-files" not found Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.422790 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-c08d-account-create-update-54vb7"] Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.424359 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c08d-account-create-update-54vb7" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.431327 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.443704 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-c08d-account-create-update-54vb7"] Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.450604 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sb6wl\" (UniqueName: \"kubernetes.io/projected/4d815209-521c-4c7f-a026-18899832459f-kube-api-access-sb6wl\") pod \"glance-db-create-fs9rr\" (UID: \"4d815209-521c-4c7f-a026-18899832459f\") " pod="openstack/glance-db-create-fs9rr" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.450828 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d815209-521c-4c7f-a026-18899832459f-operator-scripts\") pod \"glance-db-create-fs9rr\" (UID: \"4d815209-521c-4c7f-a026-18899832459f\") " pod="openstack/glance-db-create-fs9rr" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.553943 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d815209-521c-4c7f-a026-18899832459f-operator-scripts\") pod \"glance-db-create-fs9rr\" (UID: \"4d815209-521c-4c7f-a026-18899832459f\") " pod="openstack/glance-db-create-fs9rr" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.554111 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6bll\" (UniqueName: \"kubernetes.io/projected/333ba21a-82a9-4f39-9ec5-afb7ae2ccf60-kube-api-access-v6bll\") pod \"glance-c08d-account-create-update-54vb7\" (UID: \"333ba21a-82a9-4f39-9ec5-afb7ae2ccf60\") " pod="openstack/glance-c08d-account-create-update-54vb7" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.554185 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sb6wl\" (UniqueName: \"kubernetes.io/projected/4d815209-521c-4c7f-a026-18899832459f-kube-api-access-sb6wl\") pod \"glance-db-create-fs9rr\" (UID: \"4d815209-521c-4c7f-a026-18899832459f\") " pod="openstack/glance-db-create-fs9rr" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.554231 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/333ba21a-82a9-4f39-9ec5-afb7ae2ccf60-operator-scripts\") pod \"glance-c08d-account-create-update-54vb7\" (UID: \"333ba21a-82a9-4f39-9ec5-afb7ae2ccf60\") " pod="openstack/glance-c08d-account-create-update-54vb7" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.554944 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d815209-521c-4c7f-a026-18899832459f-operator-scripts\") pod \"glance-db-create-fs9rr\" (UID: \"4d815209-521c-4c7f-a026-18899832459f\") " pod="openstack/glance-db-create-fs9rr" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.577863 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sb6wl\" (UniqueName: \"kubernetes.io/projected/4d815209-521c-4c7f-a026-18899832459f-kube-api-access-sb6wl\") pod \"glance-db-create-fs9rr\" (UID: \"4d815209-521c-4c7f-a026-18899832459f\") " pod="openstack/glance-db-create-fs9rr" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.637625 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-fs9rr" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.656945 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6bll\" (UniqueName: \"kubernetes.io/projected/333ba21a-82a9-4f39-9ec5-afb7ae2ccf60-kube-api-access-v6bll\") pod \"glance-c08d-account-create-update-54vb7\" (UID: \"333ba21a-82a9-4f39-9ec5-afb7ae2ccf60\") " pod="openstack/glance-c08d-account-create-update-54vb7" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.657096 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/333ba21a-82a9-4f39-9ec5-afb7ae2ccf60-operator-scripts\") pod \"glance-c08d-account-create-update-54vb7\" (UID: \"333ba21a-82a9-4f39-9ec5-afb7ae2ccf60\") " pod="openstack/glance-c08d-account-create-update-54vb7" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.657872 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/333ba21a-82a9-4f39-9ec5-afb7ae2ccf60-operator-scripts\") pod \"glance-c08d-account-create-update-54vb7\" (UID: \"333ba21a-82a9-4f39-9ec5-afb7ae2ccf60\") " pod="openstack/glance-c08d-account-create-update-54vb7" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.680592 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6bll\" (UniqueName: \"kubernetes.io/projected/333ba21a-82a9-4f39-9ec5-afb7ae2ccf60-kube-api-access-v6bll\") pod \"glance-c08d-account-create-update-54vb7\" (UID: \"333ba21a-82a9-4f39-9ec5-afb7ae2ccf60\") " pod="openstack/glance-c08d-account-create-update-54vb7" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.733034 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-tkqzb" event={"ID":"da816a64-160f-4a5c-b945-b8fa467208d6","Type":"ContainerStarted","Data":"562bf68215847fe83f99073d009b1ee89904610f3d702b2ab282baecbca78fb2"} Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.734037 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.736960 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-dllls" event={"ID":"642adfd1-e253-4e0a-80e6-860a9f5c1ae1","Type":"ContainerStarted","Data":"befd58899320f920bf3b69e7a9c823d94bc5bd3a3193731f420fe312a97c42e6"} Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.738409 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.767642 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-tkqzb" podStartSLOduration=3.767603417 podStartE2EDuration="3.767603417s" podCreationTimestamp="2026-02-20 08:21:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:21:32.766242953 +0000 UTC m=+941.740737773" watchObservedRunningTime="2026-02-20 08:21:32.767603417 +0000 UTC m=+941.742098237" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.795091 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c08d-account-create-update-54vb7" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.813504 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-f44f4"] Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.828946 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-f44f4"] Feb 20 08:21:33 crc kubenswrapper[4948]: I0220 08:21:33.413871 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-c08d-account-create-update-54vb7"] Feb 20 08:21:33 crc kubenswrapper[4948]: I0220 08:21:33.528209 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-fs9rr"] Feb 20 08:21:33 crc kubenswrapper[4948]: I0220 08:21:33.744916 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be" path="/var/lib/kubelet/pods/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be/volumes" Feb 20 08:21:33 crc kubenswrapper[4948]: I0220 08:21:33.745633 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adf79be9-a443-4d51-b854-2067624fc60f" path="/var/lib/kubelet/pods/adf79be9-a443-4d51-b854-2067624fc60f/volumes" Feb 20 08:21:34 crc kubenswrapper[4948]: I0220 08:21:34.133044 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-rnvbh"] Feb 20 08:21:34 crc kubenswrapper[4948]: I0220 08:21:34.134477 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-rnvbh" Feb 20 08:21:34 crc kubenswrapper[4948]: I0220 08:21:34.136450 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Feb 20 08:21:34 crc kubenswrapper[4948]: I0220 08:21:34.152586 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-rnvbh"] Feb 20 08:21:34 crc kubenswrapper[4948]: I0220 08:21:34.296819 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92714d74-2a44-42e9-8829-d99ac773ac75-operator-scripts\") pod \"root-account-create-update-rnvbh\" (UID: \"92714d74-2a44-42e9-8829-d99ac773ac75\") " pod="openstack/root-account-create-update-rnvbh" Feb 20 08:21:34 crc kubenswrapper[4948]: I0220 08:21:34.296964 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f22vf\" (UniqueName: \"kubernetes.io/projected/92714d74-2a44-42e9-8829-d99ac773ac75-kube-api-access-f22vf\") pod \"root-account-create-update-rnvbh\" (UID: \"92714d74-2a44-42e9-8829-d99ac773ac75\") " pod="openstack/root-account-create-update-rnvbh" Feb 20 08:21:34 crc kubenswrapper[4948]: I0220 08:21:34.398446 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f22vf\" (UniqueName: \"kubernetes.io/projected/92714d74-2a44-42e9-8829-d99ac773ac75-kube-api-access-f22vf\") pod \"root-account-create-update-rnvbh\" (UID: \"92714d74-2a44-42e9-8829-d99ac773ac75\") " pod="openstack/root-account-create-update-rnvbh" Feb 20 08:21:34 crc kubenswrapper[4948]: I0220 08:21:34.398548 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:34 crc kubenswrapper[4948]: I0220 08:21:34.398571 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92714d74-2a44-42e9-8829-d99ac773ac75-operator-scripts\") pod \"root-account-create-update-rnvbh\" (UID: \"92714d74-2a44-42e9-8829-d99ac773ac75\") " pod="openstack/root-account-create-update-rnvbh" Feb 20 08:21:34 crc kubenswrapper[4948]: E0220 08:21:34.398961 4948 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 20 08:21:34 crc kubenswrapper[4948]: E0220 08:21:34.399078 4948 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 20 08:21:34 crc kubenswrapper[4948]: E0220 08:21:34.399166 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift podName:e4d60def-5dc9-460b-92ed-e3d0157982ec nodeName:}" failed. No retries permitted until 2026-02-20 08:21:38.399136453 +0000 UTC m=+947.373631313 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift") pod "swift-storage-0" (UID: "e4d60def-5dc9-460b-92ed-e3d0157982ec") : configmap "swift-ring-files" not found Feb 20 08:21:34 crc kubenswrapper[4948]: I0220 08:21:34.399252 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92714d74-2a44-42e9-8829-d99ac773ac75-operator-scripts\") pod \"root-account-create-update-rnvbh\" (UID: \"92714d74-2a44-42e9-8829-d99ac773ac75\") " pod="openstack/root-account-create-update-rnvbh" Feb 20 08:21:34 crc kubenswrapper[4948]: I0220 08:21:34.429059 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f22vf\" (UniqueName: \"kubernetes.io/projected/92714d74-2a44-42e9-8829-d99ac773ac75-kube-api-access-f22vf\") pod \"root-account-create-update-rnvbh\" (UID: \"92714d74-2a44-42e9-8829-d99ac773ac75\") " pod="openstack/root-account-create-update-rnvbh" Feb 20 08:21:34 crc kubenswrapper[4948]: I0220 08:21:34.465180 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-rnvbh" Feb 20 08:21:35 crc kubenswrapper[4948]: W0220 08:21:35.268862 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d815209_521c_4c7f_a026_18899832459f.slice/crio-83bd7271243acafa6581de499ba5b4200de0d064e497f5bdb3745d092e153305 WatchSource:0}: Error finding container 83bd7271243acafa6581de499ba5b4200de0d064e497f5bdb3745d092e153305: Status 404 returned error can't find the container with id 83bd7271243acafa6581de499ba5b4200de0d064e497f5bdb3745d092e153305 Feb 20 08:21:35 crc kubenswrapper[4948]: W0220 08:21:35.272172 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod333ba21a_82a9_4f39_9ec5_afb7ae2ccf60.slice/crio-7eca72d6b91c88e5a2e15a5744106cd7576ef511dae119df0c47c0377f312a2d WatchSource:0}: Error finding container 7eca72d6b91c88e5a2e15a5744106cd7576ef511dae119df0c47c0377f312a2d: Status 404 returned error can't find the container with id 7eca72d6b91c88e5a2e15a5744106cd7576ef511dae119df0c47c0377f312a2d Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.486269 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-eb36-account-create-update-554g5" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.531149 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-bv65d" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.562330 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gm6g6" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.571278 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6d58-account-create-update-bbx44" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.619263 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wnscr\" (UniqueName: \"kubernetes.io/projected/f917fa7f-51f9-4c8f-bee2-11529a6d44a6-kube-api-access-wnscr\") pod \"f917fa7f-51f9-4c8f-bee2-11529a6d44a6\" (UID: \"f917fa7f-51f9-4c8f-bee2-11529a6d44a6\") " Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.619337 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f917fa7f-51f9-4c8f-bee2-11529a6d44a6-operator-scripts\") pod \"f917fa7f-51f9-4c8f-bee2-11529a6d44a6\" (UID: \"f917fa7f-51f9-4c8f-bee2-11529a6d44a6\") " Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.619392 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5bf7a21-af5d-4334-bc98-bd0f9eb9035a-operator-scripts\") pod \"b5bf7a21-af5d-4334-bc98-bd0f9eb9035a\" (UID: \"b5bf7a21-af5d-4334-bc98-bd0f9eb9035a\") " Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.619437 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j4wx7\" (UniqueName: \"kubernetes.io/projected/b5bf7a21-af5d-4334-bc98-bd0f9eb9035a-kube-api-access-j4wx7\") pod \"b5bf7a21-af5d-4334-bc98-bd0f9eb9035a\" (UID: \"b5bf7a21-af5d-4334-bc98-bd0f9eb9035a\") " Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.620169 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f917fa7f-51f9-4c8f-bee2-11529a6d44a6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f917fa7f-51f9-4c8f-bee2-11529a6d44a6" (UID: "f917fa7f-51f9-4c8f-bee2-11529a6d44a6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.620180 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5bf7a21-af5d-4334-bc98-bd0f9eb9035a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b5bf7a21-af5d-4334-bc98-bd0f9eb9035a" (UID: "b5bf7a21-af5d-4334-bc98-bd0f9eb9035a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.620558 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f917fa7f-51f9-4c8f-bee2-11529a6d44a6-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.620577 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5bf7a21-af5d-4334-bc98-bd0f9eb9035a-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.625684 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f917fa7f-51f9-4c8f-bee2-11529a6d44a6-kube-api-access-wnscr" (OuterVolumeSpecName: "kube-api-access-wnscr") pod "f917fa7f-51f9-4c8f-bee2-11529a6d44a6" (UID: "f917fa7f-51f9-4c8f-bee2-11529a6d44a6"). InnerVolumeSpecName "kube-api-access-wnscr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.638662 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5bf7a21-af5d-4334-bc98-bd0f9eb9035a-kube-api-access-j4wx7" (OuterVolumeSpecName: "kube-api-access-j4wx7") pod "b5bf7a21-af5d-4334-bc98-bd0f9eb9035a" (UID: "b5bf7a21-af5d-4334-bc98-bd0f9eb9035a"). InnerVolumeSpecName "kube-api-access-j4wx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.721594 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kqvl\" (UniqueName: \"kubernetes.io/projected/e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb-kube-api-access-2kqvl\") pod \"e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb\" (UID: \"e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb\") " Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.721670 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52238644-5964-427e-8b77-9ca228d031db-operator-scripts\") pod \"52238644-5964-427e-8b77-9ca228d031db\" (UID: \"52238644-5964-427e-8b77-9ca228d031db\") " Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.721789 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89k8g\" (UniqueName: \"kubernetes.io/projected/52238644-5964-427e-8b77-9ca228d031db-kube-api-access-89k8g\") pod \"52238644-5964-427e-8b77-9ca228d031db\" (UID: \"52238644-5964-427e-8b77-9ca228d031db\") " Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.721826 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb-operator-scripts\") pod \"e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb\" (UID: \"e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb\") " Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.722290 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wnscr\" (UniqueName: \"kubernetes.io/projected/f917fa7f-51f9-4c8f-bee2-11529a6d44a6-kube-api-access-wnscr\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.722310 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j4wx7\" (UniqueName: \"kubernetes.io/projected/b5bf7a21-af5d-4334-bc98-bd0f9eb9035a-kube-api-access-j4wx7\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.722615 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb" (UID: "e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.723246 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52238644-5964-427e-8b77-9ca228d031db-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "52238644-5964-427e-8b77-9ca228d031db" (UID: "52238644-5964-427e-8b77-9ca228d031db"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.733790 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb-kube-api-access-2kqvl" (OuterVolumeSpecName: "kube-api-access-2kqvl") pod "e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb" (UID: "e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb"). InnerVolumeSpecName "kube-api-access-2kqvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.734170 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52238644-5964-427e-8b77-9ca228d031db-kube-api-access-89k8g" (OuterVolumeSpecName: "kube-api-access-89k8g") pod "52238644-5964-427e-8b77-9ca228d031db" (UID: "52238644-5964-427e-8b77-9ca228d031db"). InnerVolumeSpecName "kube-api-access-89k8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.770308 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-rnvbh"] Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.774110 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-fs9rr" event={"ID":"4d815209-521c-4c7f-a026-18899832459f","Type":"ContainerStarted","Data":"8e0c0f098f3973e04052c6dc5c4da1ff436370579d9b1ea6857d9e0c6bed683e"} Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.774153 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-fs9rr" event={"ID":"4d815209-521c-4c7f-a026-18899832459f","Type":"ContainerStarted","Data":"83bd7271243acafa6581de499ba5b4200de0d064e497f5bdb3745d092e153305"} Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.788377 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gm6g6" event={"ID":"52238644-5964-427e-8b77-9ca228d031db","Type":"ContainerDied","Data":"79852731770fecb861112fde754b01ca4c49d355a45bb283f9e41b055faf6747"} Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.789625 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79852731770fecb861112fde754b01ca4c49d355a45bb283f9e41b055faf6747" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.788499 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gm6g6" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.795131 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-eb36-account-create-update-554g5" event={"ID":"b5bf7a21-af5d-4334-bc98-bd0f9eb9035a","Type":"ContainerDied","Data":"25d97fb00e18b6ce84c602bbc6e333780c0537c4cab9cb2b3ce67336516b898c"} Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.795178 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="25d97fb00e18b6ce84c602bbc6e333780c0537c4cab9cb2b3ce67336516b898c" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.795417 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-eb36-account-create-update-554g5" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.798595 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6d58-account-create-update-bbx44" event={"ID":"e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb","Type":"ContainerDied","Data":"fe0235453e94ed6b67ebdad926c4872aff88ca5acd235d564a2517d91ee0dafd"} Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.798628 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe0235453e94ed6b67ebdad926c4872aff88ca5acd235d564a2517d91ee0dafd" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.798707 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6d58-account-create-update-bbx44" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.801370 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-c08d-account-create-update-54vb7" event={"ID":"333ba21a-82a9-4f39-9ec5-afb7ae2ccf60","Type":"ContainerStarted","Data":"67a7241e5369f13541e0b84997b90ee0fd96a87f376c9c419e0e637fa48068da"} Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.801441 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-c08d-account-create-update-54vb7" event={"ID":"333ba21a-82a9-4f39-9ec5-afb7ae2ccf60","Type":"ContainerStarted","Data":"7eca72d6b91c88e5a2e15a5744106cd7576ef511dae119df0c47c0377f312a2d"} Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.805498 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-dllls" event={"ID":"642adfd1-e253-4e0a-80e6-860a9f5c1ae1","Type":"ContainerStarted","Data":"8fbeb28cef63fcfd89707879381c2e4ed399e4426a2e6226acf2ce1d9b7e8e2c"} Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.807132 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-fs9rr" podStartSLOduration=3.807113914 podStartE2EDuration="3.807113914s" podCreationTimestamp="2026-02-20 08:21:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:21:35.790878413 +0000 UTC m=+944.765373253" watchObservedRunningTime="2026-02-20 08:21:35.807113914 +0000 UTC m=+944.781608734" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.808223 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-bv65d" event={"ID":"f917fa7f-51f9-4c8f-bee2-11529a6d44a6","Type":"ContainerDied","Data":"7f25c2e14bdc30bc83d90e8cf3dae79ddb8cda1ee4900ffa7eb5001c2eb0fd98"} Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.808253 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f25c2e14bdc30bc83d90e8cf3dae79ddb8cda1ee4900ffa7eb5001c2eb0fd98" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.808453 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-bv65d" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.820934 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-c08d-account-create-update-54vb7" podStartSLOduration=3.8209183859999998 podStartE2EDuration="3.820918386s" podCreationTimestamp="2026-02-20 08:21:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:21:35.820175058 +0000 UTC m=+944.794669878" watchObservedRunningTime="2026-02-20 08:21:35.820918386 +0000 UTC m=+944.795413206" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.823534 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kqvl\" (UniqueName: \"kubernetes.io/projected/e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb-kube-api-access-2kqvl\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.823557 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52238644-5964-427e-8b77-9ca228d031db-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.823567 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89k8g\" (UniqueName: \"kubernetes.io/projected/52238644-5964-427e-8b77-9ca228d031db-kube-api-access-89k8g\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.823576 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.840933 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-dllls" podStartSLOduration=1.426575126 podStartE2EDuration="4.840917511s" podCreationTimestamp="2026-02-20 08:21:31 +0000 UTC" firstStartedPulling="2026-02-20 08:21:31.908838665 +0000 UTC m=+940.883333495" lastFinishedPulling="2026-02-20 08:21:35.32318106 +0000 UTC m=+944.297675880" observedRunningTime="2026-02-20 08:21:35.83885855 +0000 UTC m=+944.813353360" watchObservedRunningTime="2026-02-20 08:21:35.840917511 +0000 UTC m=+944.815412331" Feb 20 08:21:36 crc kubenswrapper[4948]: E0220 08:21:36.382294 4948 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73b84bb7_f594_4823_ac03_40fdac6ee177.slice/crio-conmon-d4b1f06d6e332e362b758285a5271800ff6a5cc95aee58142ef4093fc832e364.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod92714d74_2a44_42e9_8829_d99ac773ac75.slice/crio-5dfd385e001fae733c595da928ba9b46f6e3cc2e4d358e37840e650d21ce855a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73b84bb7_f594_4823_ac03_40fdac6ee177.slice/crio-d4b1f06d6e332e362b758285a5271800ff6a5cc95aee58142ef4093fc832e364.scope\": RecentStats: unable to find data in memory cache]" Feb 20 08:21:36 crc kubenswrapper[4948]: I0220 08:21:36.820426 4948 generic.go:334] "Generic (PLEG): container finished" podID="4d815209-521c-4c7f-a026-18899832459f" containerID="8e0c0f098f3973e04052c6dc5c4da1ff436370579d9b1ea6857d9e0c6bed683e" exitCode=0 Feb 20 08:21:36 crc kubenswrapper[4948]: I0220 08:21:36.820522 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-fs9rr" event={"ID":"4d815209-521c-4c7f-a026-18899832459f","Type":"ContainerDied","Data":"8e0c0f098f3973e04052c6dc5c4da1ff436370579d9b1ea6857d9e0c6bed683e"} Feb 20 08:21:36 crc kubenswrapper[4948]: I0220 08:21:36.824164 4948 generic.go:334] "Generic (PLEG): container finished" podID="73b84bb7-f594-4823-ac03-40fdac6ee177" containerID="d4b1f06d6e332e362b758285a5271800ff6a5cc95aee58142ef4093fc832e364" exitCode=0 Feb 20 08:21:36 crc kubenswrapper[4948]: I0220 08:21:36.824245 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"73b84bb7-f594-4823-ac03-40fdac6ee177","Type":"ContainerDied","Data":"d4b1f06d6e332e362b758285a5271800ff6a5cc95aee58142ef4093fc832e364"} Feb 20 08:21:36 crc kubenswrapper[4948]: I0220 08:21:36.829361 4948 generic.go:334] "Generic (PLEG): container finished" podID="92714d74-2a44-42e9-8829-d99ac773ac75" containerID="5dfd385e001fae733c595da928ba9b46f6e3cc2e4d358e37840e650d21ce855a" exitCode=0 Feb 20 08:21:36 crc kubenswrapper[4948]: I0220 08:21:36.829453 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-rnvbh" event={"ID":"92714d74-2a44-42e9-8829-d99ac773ac75","Type":"ContainerDied","Data":"5dfd385e001fae733c595da928ba9b46f6e3cc2e4d358e37840e650d21ce855a"} Feb 20 08:21:36 crc kubenswrapper[4948]: I0220 08:21:36.829488 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-rnvbh" event={"ID":"92714d74-2a44-42e9-8829-d99ac773ac75","Type":"ContainerStarted","Data":"5cee8fe5e9ab322142fa63fb214c8bbff5f35bb77dce1e1e2cae0a5fb25cc1c8"} Feb 20 08:21:36 crc kubenswrapper[4948]: I0220 08:21:36.832106 4948 generic.go:334] "Generic (PLEG): container finished" podID="333ba21a-82a9-4f39-9ec5-afb7ae2ccf60" containerID="67a7241e5369f13541e0b84997b90ee0fd96a87f376c9c419e0e637fa48068da" exitCode=0 Feb 20 08:21:36 crc kubenswrapper[4948]: I0220 08:21:36.832621 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-c08d-account-create-update-54vb7" event={"ID":"333ba21a-82a9-4f39-9ec5-afb7ae2ccf60","Type":"ContainerDied","Data":"67a7241e5369f13541e0b84997b90ee0fd96a87f376c9c419e0e637fa48068da"} Feb 20 08:21:37 crc kubenswrapper[4948]: I0220 08:21:37.845491 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"73b84bb7-f594-4823-ac03-40fdac6ee177","Type":"ContainerStarted","Data":"f2ef52eaf8fede6026f04fb9c9c143db4ac26c52cfdab80a2a52811b838b1e85"} Feb 20 08:21:37 crc kubenswrapper[4948]: I0220 08:21:37.846049 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 20 08:21:37 crc kubenswrapper[4948]: I0220 08:21:37.848108 4948 generic.go:334] "Generic (PLEG): container finished" podID="790894c9-053c-497c-955d-ce7519111dd6" containerID="77b026b61409d73dee6a89b7f49bf1a86167df7972db0e3eb3f4cbb10df48242" exitCode=0 Feb 20 08:21:37 crc kubenswrapper[4948]: I0220 08:21:37.848150 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"790894c9-053c-497c-955d-ce7519111dd6","Type":"ContainerDied","Data":"77b026b61409d73dee6a89b7f49bf1a86167df7972db0e3eb3f4cbb10df48242"} Feb 20 08:21:37 crc kubenswrapper[4948]: I0220 08:21:37.877493 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.429138858 podStartE2EDuration="55.8774688s" podCreationTimestamp="2026-02-20 08:20:42 +0000 UTC" firstStartedPulling="2026-02-20 08:20:44.406206929 +0000 UTC m=+893.380701749" lastFinishedPulling="2026-02-20 08:21:02.854536861 +0000 UTC m=+911.829031691" observedRunningTime="2026-02-20 08:21:37.873259996 +0000 UTC m=+946.847754816" watchObservedRunningTime="2026-02-20 08:21:37.8774688 +0000 UTC m=+946.851963640" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.024852 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.026368 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.026605 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.027849 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"df8e056c06ac241c46a927c4e3f5a28f39d042bf13a8c672499d528786dbbbb0"} pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.027920 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" containerID="cri-o://df8e056c06ac241c46a927c4e3f5a28f39d042bf13a8c672499d528786dbbbb0" gracePeriod=600 Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.373964 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c08d-account-create-update-54vb7" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.391079 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-fs9rr" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.423803 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-rnvbh" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.491058 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6wl\" (UniqueName: \"kubernetes.io/projected/4d815209-521c-4c7f-a026-18899832459f-kube-api-access-sb6wl\") pod \"4d815209-521c-4c7f-a026-18899832459f\" (UID: \"4d815209-521c-4c7f-a026-18899832459f\") " Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.491234 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/333ba21a-82a9-4f39-9ec5-afb7ae2ccf60-operator-scripts\") pod \"333ba21a-82a9-4f39-9ec5-afb7ae2ccf60\" (UID: \"333ba21a-82a9-4f39-9ec5-afb7ae2ccf60\") " Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.491355 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92714d74-2a44-42e9-8829-d99ac773ac75-operator-scripts\") pod \"92714d74-2a44-42e9-8829-d99ac773ac75\" (UID: \"92714d74-2a44-42e9-8829-d99ac773ac75\") " Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.491490 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d815209-521c-4c7f-a026-18899832459f-operator-scripts\") pod \"4d815209-521c-4c7f-a026-18899832459f\" (UID: \"4d815209-521c-4c7f-a026-18899832459f\") " Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.491515 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6bll\" (UniqueName: \"kubernetes.io/projected/333ba21a-82a9-4f39-9ec5-afb7ae2ccf60-kube-api-access-v6bll\") pod \"333ba21a-82a9-4f39-9ec5-afb7ae2ccf60\" (UID: \"333ba21a-82a9-4f39-9ec5-afb7ae2ccf60\") " Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.491603 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f22vf\" (UniqueName: \"kubernetes.io/projected/92714d74-2a44-42e9-8829-d99ac773ac75-kube-api-access-f22vf\") pod \"92714d74-2a44-42e9-8829-d99ac773ac75\" (UID: \"92714d74-2a44-42e9-8829-d99ac773ac75\") " Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.491831 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:38 crc kubenswrapper[4948]: E0220 08:21:38.492433 4948 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 20 08:21:38 crc kubenswrapper[4948]: E0220 08:21:38.492519 4948 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 20 08:21:38 crc kubenswrapper[4948]: E0220 08:21:38.492606 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift podName:e4d60def-5dc9-460b-92ed-e3d0157982ec nodeName:}" failed. No retries permitted until 2026-02-20 08:21:46.492591302 +0000 UTC m=+955.467086122 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift") pod "swift-storage-0" (UID: "e4d60def-5dc9-460b-92ed-e3d0157982ec") : configmap "swift-ring-files" not found Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.495410 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/333ba21a-82a9-4f39-9ec5-afb7ae2ccf60-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "333ba21a-82a9-4f39-9ec5-afb7ae2ccf60" (UID: "333ba21a-82a9-4f39-9ec5-afb7ae2ccf60"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.495480 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d815209-521c-4c7f-a026-18899832459f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4d815209-521c-4c7f-a026-18899832459f" (UID: "4d815209-521c-4c7f-a026-18899832459f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.495617 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92714d74-2a44-42e9-8829-d99ac773ac75-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "92714d74-2a44-42e9-8829-d99ac773ac75" (UID: "92714d74-2a44-42e9-8829-d99ac773ac75"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.501133 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d815209-521c-4c7f-a026-18899832459f-kube-api-access-sb6wl" (OuterVolumeSpecName: "kube-api-access-sb6wl") pod "4d815209-521c-4c7f-a026-18899832459f" (UID: "4d815209-521c-4c7f-a026-18899832459f"). InnerVolumeSpecName "kube-api-access-sb6wl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.501170 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/333ba21a-82a9-4f39-9ec5-afb7ae2ccf60-kube-api-access-v6bll" (OuterVolumeSpecName: "kube-api-access-v6bll") pod "333ba21a-82a9-4f39-9ec5-afb7ae2ccf60" (UID: "333ba21a-82a9-4f39-9ec5-afb7ae2ccf60"). InnerVolumeSpecName "kube-api-access-v6bll". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.501187 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92714d74-2a44-42e9-8829-d99ac773ac75-kube-api-access-f22vf" (OuterVolumeSpecName: "kube-api-access-f22vf") pod "92714d74-2a44-42e9-8829-d99ac773ac75" (UID: "92714d74-2a44-42e9-8829-d99ac773ac75"). InnerVolumeSpecName "kube-api-access-f22vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.593843 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f22vf\" (UniqueName: \"kubernetes.io/projected/92714d74-2a44-42e9-8829-d99ac773ac75-kube-api-access-f22vf\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.594188 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6wl\" (UniqueName: \"kubernetes.io/projected/4d815209-521c-4c7f-a026-18899832459f-kube-api-access-sb6wl\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.594198 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/333ba21a-82a9-4f39-9ec5-afb7ae2ccf60-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.594207 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92714d74-2a44-42e9-8829-d99ac773ac75-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.594216 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d815209-521c-4c7f-a026-18899832459f-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.594224 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6bll\" (UniqueName: \"kubernetes.io/projected/333ba21a-82a9-4f39-9ec5-afb7ae2ccf60-kube-api-access-v6bll\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.857240 4948 generic.go:334] "Generic (PLEG): container finished" podID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerID="df8e056c06ac241c46a927c4e3f5a28f39d042bf13a8c672499d528786dbbbb0" exitCode=0 Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.857271 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerDied","Data":"df8e056c06ac241c46a927c4e3f5a28f39d042bf13a8c672499d528786dbbbb0"} Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.858376 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"8f0b8a806330c630468951158113ee0ecaad5dbddb81b09c063752a17f21282c"} Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.858416 4948 scope.go:117] "RemoveContainer" containerID="b5405a03f834a339ac758e6efb1e31a8c094b3f6b973ea8540eefc716ba7bf7f" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.858594 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-c08d-account-create-update-54vb7" event={"ID":"333ba21a-82a9-4f39-9ec5-afb7ae2ccf60","Type":"ContainerDied","Data":"7eca72d6b91c88e5a2e15a5744106cd7576ef511dae119df0c47c0377f312a2d"} Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.858637 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c08d-account-create-update-54vb7" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.858639 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7eca72d6b91c88e5a2e15a5744106cd7576ef511dae119df0c47c0377f312a2d" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.872919 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-rnvbh" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.873085 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-rnvbh" event={"ID":"92714d74-2a44-42e9-8829-d99ac773ac75","Type":"ContainerDied","Data":"5cee8fe5e9ab322142fa63fb214c8bbff5f35bb77dce1e1e2cae0a5fb25cc1c8"} Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.873134 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5cee8fe5e9ab322142fa63fb214c8bbff5f35bb77dce1e1e2cae0a5fb25cc1c8" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.874956 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"790894c9-053c-497c-955d-ce7519111dd6","Type":"ContainerStarted","Data":"ba12ca5de96a23b6d0de36b28eae9e3fb2a92d4464edd4582c10c73ae804254a"} Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.875502 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.877185 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-fs9rr" event={"ID":"4d815209-521c-4c7f-a026-18899832459f","Type":"ContainerDied","Data":"83bd7271243acafa6581de499ba5b4200de0d064e497f5bdb3745d092e153305"} Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.877217 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83bd7271243acafa6581de499ba5b4200de0d064e497f5bdb3745d092e153305" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.877218 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-fs9rr" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.915250 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=55.915236221 podStartE2EDuration="55.915236221s" podCreationTimestamp="2026-02-20 08:20:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:21:38.914516603 +0000 UTC m=+947.889011453" watchObservedRunningTime="2026-02-20 08:21:38.915236221 +0000 UTC m=+947.889731041" Feb 20 08:21:39 crc kubenswrapper[4948]: I0220 08:21:39.849020 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:39 crc kubenswrapper[4948]: I0220 08:21:39.912950 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-hmhzl"] Feb 20 08:21:39 crc kubenswrapper[4948]: I0220 08:21:39.913251 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" podUID="33a06f41-ea8b-4138-8048-5425cd04f015" containerName="dnsmasq-dns" containerID="cri-o://489a052e4c0fb8182ef93444ec3a663e93256a532f88c70856200cb8dc8b254f" gracePeriod=10 Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.378172 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.422477 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-config\") pod \"33a06f41-ea8b-4138-8048-5425cd04f015\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.422529 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-dns-svc\") pod \"33a06f41-ea8b-4138-8048-5425cd04f015\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.422595 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-ovsdbserver-nb\") pod \"33a06f41-ea8b-4138-8048-5425cd04f015\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.422744 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2twp\" (UniqueName: \"kubernetes.io/projected/33a06f41-ea8b-4138-8048-5425cd04f015-kube-api-access-c2twp\") pod \"33a06f41-ea8b-4138-8048-5425cd04f015\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.422910 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-ovsdbserver-sb\") pod \"33a06f41-ea8b-4138-8048-5425cd04f015\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.439889 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33a06f41-ea8b-4138-8048-5425cd04f015-kube-api-access-c2twp" (OuterVolumeSpecName: "kube-api-access-c2twp") pod "33a06f41-ea8b-4138-8048-5425cd04f015" (UID: "33a06f41-ea8b-4138-8048-5425cd04f015"). InnerVolumeSpecName "kube-api-access-c2twp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.458863 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-config" (OuterVolumeSpecName: "config") pod "33a06f41-ea8b-4138-8048-5425cd04f015" (UID: "33a06f41-ea8b-4138-8048-5425cd04f015"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.464527 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "33a06f41-ea8b-4138-8048-5425cd04f015" (UID: "33a06f41-ea8b-4138-8048-5425cd04f015"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.465764 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "33a06f41-ea8b-4138-8048-5425cd04f015" (UID: "33a06f41-ea8b-4138-8048-5425cd04f015"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.467496 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "33a06f41-ea8b-4138-8048-5425cd04f015" (UID: "33a06f41-ea8b-4138-8048-5425cd04f015"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.525394 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.525420 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.525429 4948 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.525438 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.525446 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2twp\" (UniqueName: \"kubernetes.io/projected/33a06f41-ea8b-4138-8048-5425cd04f015-kube-api-access-c2twp\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.539324 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-rnvbh"] Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.544293 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-rnvbh"] Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.896203 4948 generic.go:334] "Generic (PLEG): container finished" podID="33a06f41-ea8b-4138-8048-5425cd04f015" containerID="489a052e4c0fb8182ef93444ec3a663e93256a532f88c70856200cb8dc8b254f" exitCode=0 Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.896249 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" event={"ID":"33a06f41-ea8b-4138-8048-5425cd04f015","Type":"ContainerDied","Data":"489a052e4c0fb8182ef93444ec3a663e93256a532f88c70856200cb8dc8b254f"} Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.896277 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" event={"ID":"33a06f41-ea8b-4138-8048-5425cd04f015","Type":"ContainerDied","Data":"1f7e5587a47afc9ffa01b7b40dc0afdae69e9922140a4a9e4162b6a2a7417196"} Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.896296 4948 scope.go:117] "RemoveContainer" containerID="489a052e4c0fb8182ef93444ec3a663e93256a532f88c70856200cb8dc8b254f" Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.896395 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.953783 4948 scope.go:117] "RemoveContainer" containerID="664c1ad747d8bd0a828811f06f4556fb8972cb9b47fe4b7d90fa3826c1e33b4a" Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.956098 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-hmhzl"] Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.966526 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-hmhzl"] Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.979825 4948 scope.go:117] "RemoveContainer" containerID="489a052e4c0fb8182ef93444ec3a663e93256a532f88c70856200cb8dc8b254f" Feb 20 08:21:40 crc kubenswrapper[4948]: E0220 08:21:40.980302 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"489a052e4c0fb8182ef93444ec3a663e93256a532f88c70856200cb8dc8b254f\": container with ID starting with 489a052e4c0fb8182ef93444ec3a663e93256a532f88c70856200cb8dc8b254f not found: ID does not exist" containerID="489a052e4c0fb8182ef93444ec3a663e93256a532f88c70856200cb8dc8b254f" Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.980345 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"489a052e4c0fb8182ef93444ec3a663e93256a532f88c70856200cb8dc8b254f"} err="failed to get container status \"489a052e4c0fb8182ef93444ec3a663e93256a532f88c70856200cb8dc8b254f\": rpc error: code = NotFound desc = could not find container \"489a052e4c0fb8182ef93444ec3a663e93256a532f88c70856200cb8dc8b254f\": container with ID starting with 489a052e4c0fb8182ef93444ec3a663e93256a532f88c70856200cb8dc8b254f not found: ID does not exist" Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.980370 4948 scope.go:117] "RemoveContainer" containerID="664c1ad747d8bd0a828811f06f4556fb8972cb9b47fe4b7d90fa3826c1e33b4a" Feb 20 08:21:40 crc kubenswrapper[4948]: E0220 08:21:40.980705 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"664c1ad747d8bd0a828811f06f4556fb8972cb9b47fe4b7d90fa3826c1e33b4a\": container with ID starting with 664c1ad747d8bd0a828811f06f4556fb8972cb9b47fe4b7d90fa3826c1e33b4a not found: ID does not exist" containerID="664c1ad747d8bd0a828811f06f4556fb8972cb9b47fe4b7d90fa3826c1e33b4a" Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.980731 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"664c1ad747d8bd0a828811f06f4556fb8972cb9b47fe4b7d90fa3826c1e33b4a"} err="failed to get container status \"664c1ad747d8bd0a828811f06f4556fb8972cb9b47fe4b7d90fa3826c1e33b4a\": rpc error: code = NotFound desc = could not find container \"664c1ad747d8bd0a828811f06f4556fb8972cb9b47fe4b7d90fa3826c1e33b4a\": container with ID starting with 664c1ad747d8bd0a828811f06f4556fb8972cb9b47fe4b7d90fa3826c1e33b4a not found: ID does not exist" Feb 20 08:21:41 crc kubenswrapper[4948]: I0220 08:21:41.464209 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Feb 20 08:21:41 crc kubenswrapper[4948]: I0220 08:21:41.731513 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33a06f41-ea8b-4138-8048-5425cd04f015" path="/var/lib/kubelet/pods/33a06f41-ea8b-4138-8048-5425cd04f015/volumes" Feb 20 08:21:41 crc kubenswrapper[4948]: I0220 08:21:41.732126 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92714d74-2a44-42e9-8829-d99ac773ac75" path="/var/lib/kubelet/pods/92714d74-2a44-42e9-8829-d99ac773ac75/volumes" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.511717 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-pkgzx"] Feb 20 08:21:42 crc kubenswrapper[4948]: E0220 08:21:42.512159 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33a06f41-ea8b-4138-8048-5425cd04f015" containerName="dnsmasq-dns" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.512203 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="33a06f41-ea8b-4138-8048-5425cd04f015" containerName="dnsmasq-dns" Feb 20 08:21:42 crc kubenswrapper[4948]: E0220 08:21:42.512223 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52238644-5964-427e-8b77-9ca228d031db" containerName="mariadb-database-create" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.512231 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="52238644-5964-427e-8b77-9ca228d031db" containerName="mariadb-database-create" Feb 20 08:21:42 crc kubenswrapper[4948]: E0220 08:21:42.512248 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb" containerName="mariadb-account-create-update" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.512257 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb" containerName="mariadb-account-create-update" Feb 20 08:21:42 crc kubenswrapper[4948]: E0220 08:21:42.512272 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="333ba21a-82a9-4f39-9ec5-afb7ae2ccf60" containerName="mariadb-account-create-update" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.512279 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="333ba21a-82a9-4f39-9ec5-afb7ae2ccf60" containerName="mariadb-account-create-update" Feb 20 08:21:42 crc kubenswrapper[4948]: E0220 08:21:42.512289 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f917fa7f-51f9-4c8f-bee2-11529a6d44a6" containerName="mariadb-database-create" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.512297 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f917fa7f-51f9-4c8f-bee2-11529a6d44a6" containerName="mariadb-database-create" Feb 20 08:21:42 crc kubenswrapper[4948]: E0220 08:21:42.512311 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d815209-521c-4c7f-a026-18899832459f" containerName="mariadb-database-create" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.512319 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d815209-521c-4c7f-a026-18899832459f" containerName="mariadb-database-create" Feb 20 08:21:42 crc kubenswrapper[4948]: E0220 08:21:42.512334 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92714d74-2a44-42e9-8829-d99ac773ac75" containerName="mariadb-account-create-update" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.512342 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="92714d74-2a44-42e9-8829-d99ac773ac75" containerName="mariadb-account-create-update" Feb 20 08:21:42 crc kubenswrapper[4948]: E0220 08:21:42.512356 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33a06f41-ea8b-4138-8048-5425cd04f015" containerName="init" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.512364 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="33a06f41-ea8b-4138-8048-5425cd04f015" containerName="init" Feb 20 08:21:42 crc kubenswrapper[4948]: E0220 08:21:42.512377 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5bf7a21-af5d-4334-bc98-bd0f9eb9035a" containerName="mariadb-account-create-update" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.512386 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5bf7a21-af5d-4334-bc98-bd0f9eb9035a" containerName="mariadb-account-create-update" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.512567 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f917fa7f-51f9-4c8f-bee2-11529a6d44a6" containerName="mariadb-database-create" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.512586 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb" containerName="mariadb-account-create-update" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.512595 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5bf7a21-af5d-4334-bc98-bd0f9eb9035a" containerName="mariadb-account-create-update" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.512611 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="92714d74-2a44-42e9-8829-d99ac773ac75" containerName="mariadb-account-create-update" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.512620 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d815209-521c-4c7f-a026-18899832459f" containerName="mariadb-database-create" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.512636 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="33a06f41-ea8b-4138-8048-5425cd04f015" containerName="dnsmasq-dns" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.512648 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="52238644-5964-427e-8b77-9ca228d031db" containerName="mariadb-database-create" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.512658 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="333ba21a-82a9-4f39-9ec5-afb7ae2ccf60" containerName="mariadb-account-create-update" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.513304 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-pkgzx" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.529576 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.534610 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-bgxj8" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.543634 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-pkgzx"] Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.660512 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-db-sync-config-data\") pod \"glance-db-sync-pkgzx\" (UID: \"65735728-e734-40de-a185-9caa1db9a47d\") " pod="openstack/glance-db-sync-pkgzx" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.660571 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-combined-ca-bundle\") pod \"glance-db-sync-pkgzx\" (UID: \"65735728-e734-40de-a185-9caa1db9a47d\") " pod="openstack/glance-db-sync-pkgzx" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.660640 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-config-data\") pod \"glance-db-sync-pkgzx\" (UID: \"65735728-e734-40de-a185-9caa1db9a47d\") " pod="openstack/glance-db-sync-pkgzx" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.661378 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j589d\" (UniqueName: \"kubernetes.io/projected/65735728-e734-40de-a185-9caa1db9a47d-kube-api-access-j589d\") pod \"glance-db-sync-pkgzx\" (UID: \"65735728-e734-40de-a185-9caa1db9a47d\") " pod="openstack/glance-db-sync-pkgzx" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.762997 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-config-data\") pod \"glance-db-sync-pkgzx\" (UID: \"65735728-e734-40de-a185-9caa1db9a47d\") " pod="openstack/glance-db-sync-pkgzx" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.763080 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j589d\" (UniqueName: \"kubernetes.io/projected/65735728-e734-40de-a185-9caa1db9a47d-kube-api-access-j589d\") pod \"glance-db-sync-pkgzx\" (UID: \"65735728-e734-40de-a185-9caa1db9a47d\") " pod="openstack/glance-db-sync-pkgzx" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.763124 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-db-sync-config-data\") pod \"glance-db-sync-pkgzx\" (UID: \"65735728-e734-40de-a185-9caa1db9a47d\") " pod="openstack/glance-db-sync-pkgzx" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.763157 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-combined-ca-bundle\") pod \"glance-db-sync-pkgzx\" (UID: \"65735728-e734-40de-a185-9caa1db9a47d\") " pod="openstack/glance-db-sync-pkgzx" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.771908 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-combined-ca-bundle\") pod \"glance-db-sync-pkgzx\" (UID: \"65735728-e734-40de-a185-9caa1db9a47d\") " pod="openstack/glance-db-sync-pkgzx" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.778518 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-db-sync-config-data\") pod \"glance-db-sync-pkgzx\" (UID: \"65735728-e734-40de-a185-9caa1db9a47d\") " pod="openstack/glance-db-sync-pkgzx" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.778545 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-config-data\") pod \"glance-db-sync-pkgzx\" (UID: \"65735728-e734-40de-a185-9caa1db9a47d\") " pod="openstack/glance-db-sync-pkgzx" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.783678 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j589d\" (UniqueName: \"kubernetes.io/projected/65735728-e734-40de-a185-9caa1db9a47d-kube-api-access-j589d\") pod \"glance-db-sync-pkgzx\" (UID: \"65735728-e734-40de-a185-9caa1db9a47d\") " pod="openstack/glance-db-sync-pkgzx" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.838138 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-pkgzx" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.915019 4948 generic.go:334] "Generic (PLEG): container finished" podID="642adfd1-e253-4e0a-80e6-860a9f5c1ae1" containerID="8fbeb28cef63fcfd89707879381c2e4ed399e4426a2e6226acf2ce1d9b7e8e2c" exitCode=0 Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.915072 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-dllls" event={"ID":"642adfd1-e253-4e0a-80e6-860a9f5c1ae1","Type":"ContainerDied","Data":"8fbeb28cef63fcfd89707879381c2e4ed399e4426a2e6226acf2ce1d9b7e8e2c"} Feb 20 08:21:43 crc kubenswrapper[4948]: I0220 08:21:43.389633 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-pkgzx"] Feb 20 08:21:43 crc kubenswrapper[4948]: W0220 08:21:43.398704 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65735728_e734_40de_a185_9caa1db9a47d.slice/crio-5e99cbf3924927a30ad54b19ba1a0b9e160811b4e81053363ade5b0662fa69fa WatchSource:0}: Error finding container 5e99cbf3924927a30ad54b19ba1a0b9e160811b4e81053363ade5b0662fa69fa: Status 404 returned error can't find the container with id 5e99cbf3924927a30ad54b19ba1a0b9e160811b4e81053363ade5b0662fa69fa Feb 20 08:21:43 crc kubenswrapper[4948]: I0220 08:21:43.925186 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-pkgzx" event={"ID":"65735728-e734-40de-a185-9caa1db9a47d","Type":"ContainerStarted","Data":"5e99cbf3924927a30ad54b19ba1a0b9e160811b4e81053363ade5b0662fa69fa"} Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.231696 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-n7gxj"] Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.252319 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-n7gxj" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.261105 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.262630 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-n7gxj"] Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.287055 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.287275 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/538046ba-61c4-43ff-ad1f-a8d80d2ab075-operator-scripts\") pod \"root-account-create-update-n7gxj\" (UID: \"538046ba-61c4-43ff-ad1f-a8d80d2ab075\") " pod="openstack/root-account-create-update-n7gxj" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.287330 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxckc\" (UniqueName: \"kubernetes.io/projected/538046ba-61c4-43ff-ad1f-a8d80d2ab075-kube-api-access-jxckc\") pod \"root-account-create-update-n7gxj\" (UID: \"538046ba-61c4-43ff-ad1f-a8d80d2ab075\") " pod="openstack/root-account-create-update-n7gxj" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.389091 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-scripts\") pod \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.389132 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-combined-ca-bundle\") pod \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.389197 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-dispersionconf\") pod \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.389223 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-swiftconf\") pod \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.389299 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-ring-data-devices\") pod \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.389335 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-etc-swift\") pod \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.389362 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89jgn\" (UniqueName: \"kubernetes.io/projected/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-kube-api-access-89jgn\") pod \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.389580 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/538046ba-61c4-43ff-ad1f-a8d80d2ab075-operator-scripts\") pod \"root-account-create-update-n7gxj\" (UID: \"538046ba-61c4-43ff-ad1f-a8d80d2ab075\") " pod="openstack/root-account-create-update-n7gxj" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.389608 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxckc\" (UniqueName: \"kubernetes.io/projected/538046ba-61c4-43ff-ad1f-a8d80d2ab075-kube-api-access-jxckc\") pod \"root-account-create-update-n7gxj\" (UID: \"538046ba-61c4-43ff-ad1f-a8d80d2ab075\") " pod="openstack/root-account-create-update-n7gxj" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.390339 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "642adfd1-e253-4e0a-80e6-860a9f5c1ae1" (UID: "642adfd1-e253-4e0a-80e6-860a9f5c1ae1"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.390395 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "642adfd1-e253-4e0a-80e6-860a9f5c1ae1" (UID: "642adfd1-e253-4e0a-80e6-860a9f5c1ae1"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.390633 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/538046ba-61c4-43ff-ad1f-a8d80d2ab075-operator-scripts\") pod \"root-account-create-update-n7gxj\" (UID: \"538046ba-61c4-43ff-ad1f-a8d80d2ab075\") " pod="openstack/root-account-create-update-n7gxj" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.395504 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-kube-api-access-89jgn" (OuterVolumeSpecName: "kube-api-access-89jgn") pod "642adfd1-e253-4e0a-80e6-860a9f5c1ae1" (UID: "642adfd1-e253-4e0a-80e6-860a9f5c1ae1"). InnerVolumeSpecName "kube-api-access-89jgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.398277 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "642adfd1-e253-4e0a-80e6-860a9f5c1ae1" (UID: "642adfd1-e253-4e0a-80e6-860a9f5c1ae1"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.406491 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxckc\" (UniqueName: \"kubernetes.io/projected/538046ba-61c4-43ff-ad1f-a8d80d2ab075-kube-api-access-jxckc\") pod \"root-account-create-update-n7gxj\" (UID: \"538046ba-61c4-43ff-ad1f-a8d80d2ab075\") " pod="openstack/root-account-create-update-n7gxj" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.416814 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "642adfd1-e253-4e0a-80e6-860a9f5c1ae1" (UID: "642adfd1-e253-4e0a-80e6-860a9f5c1ae1"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.417088 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-scripts" (OuterVolumeSpecName: "scripts") pod "642adfd1-e253-4e0a-80e6-860a9f5c1ae1" (UID: "642adfd1-e253-4e0a-80e6-860a9f5c1ae1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.424927 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "642adfd1-e253-4e0a-80e6-860a9f5c1ae1" (UID: "642adfd1-e253-4e0a-80e6-860a9f5c1ae1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.490350 4948 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.490383 4948 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.490393 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89jgn\" (UniqueName: \"kubernetes.io/projected/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-kube-api-access-89jgn\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.490404 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.490413 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.490421 4948 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.490431 4948 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.598195 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-n7gxj" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.934631 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-dllls" event={"ID":"642adfd1-e253-4e0a-80e6-860a9f5c1ae1","Type":"ContainerDied","Data":"befd58899320f920bf3b69e7a9c823d94bc5bd3a3193731f420fe312a97c42e6"} Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.934669 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="befd58899320f920bf3b69e7a9c823d94bc5bd3a3193731f420fe312a97c42e6" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.934700 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:45 crc kubenswrapper[4948]: I0220 08:21:45.059680 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-n7gxj"] Feb 20 08:21:45 crc kubenswrapper[4948]: W0220 08:21:45.064233 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod538046ba_61c4_43ff_ad1f_a8d80d2ab075.slice/crio-61b1645ad6199f5a1d388077d4b580c92f8d45bff4fda65d7f5262e0a615940a WatchSource:0}: Error finding container 61b1645ad6199f5a1d388077d4b580c92f8d45bff4fda65d7f5262e0a615940a: Status 404 returned error can't find the container with id 61b1645ad6199f5a1d388077d4b580c92f8d45bff4fda65d7f5262e0a615940a Feb 20 08:21:45 crc kubenswrapper[4948]: I0220 08:21:45.944143 4948 generic.go:334] "Generic (PLEG): container finished" podID="538046ba-61c4-43ff-ad1f-a8d80d2ab075" containerID="c4af4c0a8310dde22c0962922d9e1a2de48ad4ce8cecfbf27d7c0a78b93a2059" exitCode=0 Feb 20 08:21:45 crc kubenswrapper[4948]: I0220 08:21:45.944280 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-n7gxj" event={"ID":"538046ba-61c4-43ff-ad1f-a8d80d2ab075","Type":"ContainerDied","Data":"c4af4c0a8310dde22c0962922d9e1a2de48ad4ce8cecfbf27d7c0a78b93a2059"} Feb 20 08:21:45 crc kubenswrapper[4948]: I0220 08:21:45.944430 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-n7gxj" event={"ID":"538046ba-61c4-43ff-ad1f-a8d80d2ab075","Type":"ContainerStarted","Data":"61b1645ad6199f5a1d388077d4b580c92f8d45bff4fda65d7f5262e0a615940a"} Feb 20 08:21:46 crc kubenswrapper[4948]: I0220 08:21:46.543774 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:46 crc kubenswrapper[4948]: I0220 08:21:46.553421 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:46 crc kubenswrapper[4948]: I0220 08:21:46.581332 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.092253 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.208243 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-mkpng" podUID="b80cb988-de2a-4e65-9161-a0af0561c754" containerName="ovn-controller" probeResult="failure" output=< Feb 20 08:21:47 crc kubenswrapper[4948]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Feb 20 08:21:47 crc kubenswrapper[4948]: > Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.238603 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.244142 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.288632 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-n7gxj" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.476407 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxckc\" (UniqueName: \"kubernetes.io/projected/538046ba-61c4-43ff-ad1f-a8d80d2ab075-kube-api-access-jxckc\") pod \"538046ba-61c4-43ff-ad1f-a8d80d2ab075\" (UID: \"538046ba-61c4-43ff-ad1f-a8d80d2ab075\") " Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.476554 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/538046ba-61c4-43ff-ad1f-a8d80d2ab075-operator-scripts\") pod \"538046ba-61c4-43ff-ad1f-a8d80d2ab075\" (UID: \"538046ba-61c4-43ff-ad1f-a8d80d2ab075\") " Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.477366 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/538046ba-61c4-43ff-ad1f-a8d80d2ab075-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "538046ba-61c4-43ff-ad1f-a8d80d2ab075" (UID: "538046ba-61c4-43ff-ad1f-a8d80d2ab075"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.484300 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/538046ba-61c4-43ff-ad1f-a8d80d2ab075-kube-api-access-jxckc" (OuterVolumeSpecName: "kube-api-access-jxckc") pod "538046ba-61c4-43ff-ad1f-a8d80d2ab075" (UID: "538046ba-61c4-43ff-ad1f-a8d80d2ab075"). InnerVolumeSpecName "kube-api-access-jxckc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.491334 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-mkpng-config-g4w5d"] Feb 20 08:21:47 crc kubenswrapper[4948]: E0220 08:21:47.491692 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="642adfd1-e253-4e0a-80e6-860a9f5c1ae1" containerName="swift-ring-rebalance" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.491705 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="642adfd1-e253-4e0a-80e6-860a9f5c1ae1" containerName="swift-ring-rebalance" Feb 20 08:21:47 crc kubenswrapper[4948]: E0220 08:21:47.491717 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="538046ba-61c4-43ff-ad1f-a8d80d2ab075" containerName="mariadb-account-create-update" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.491722 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="538046ba-61c4-43ff-ad1f-a8d80d2ab075" containerName="mariadb-account-create-update" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.493396 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="642adfd1-e253-4e0a-80e6-860a9f5c1ae1" containerName="swift-ring-rebalance" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.493417 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="538046ba-61c4-43ff-ad1f-a8d80d2ab075" containerName="mariadb-account-create-update" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.494022 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.497952 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.557995 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-mkpng-config-g4w5d"] Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.578973 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-run\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.579076 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-run-ovn\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.579096 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fa672fa0-b98b-477c-a34c-ce996ce6592d-scripts\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.579117 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-log-ovn\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.579197 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8667m\" (UniqueName: \"kubernetes.io/projected/fa672fa0-b98b-477c-a34c-ce996ce6592d-kube-api-access-8667m\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.579322 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/fa672fa0-b98b-477c-a34c-ce996ce6592d-additional-scripts\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.579517 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxckc\" (UniqueName: \"kubernetes.io/projected/538046ba-61c4-43ff-ad1f-a8d80d2ab075-kube-api-access-jxckc\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.579542 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/538046ba-61c4-43ff-ad1f-a8d80d2ab075-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.681544 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-run\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.681706 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-run-ovn\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.681730 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fa672fa0-b98b-477c-a34c-ce996ce6592d-scripts\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.681762 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-log-ovn\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.681758 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-run\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.681832 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-run-ovn\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.681922 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8667m\" (UniqueName: \"kubernetes.io/projected/fa672fa0-b98b-477c-a34c-ce996ce6592d-kube-api-access-8667m\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.681971 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/fa672fa0-b98b-477c-a34c-ce996ce6592d-additional-scripts\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.683039 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/fa672fa0-b98b-477c-a34c-ce996ce6592d-additional-scripts\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.683098 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-log-ovn\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.684140 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fa672fa0-b98b-477c-a34c-ce996ce6592d-scripts\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.700230 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8667m\" (UniqueName: \"kubernetes.io/projected/fa672fa0-b98b-477c-a34c-ce996ce6592d-kube-api-access-8667m\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.900163 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.963791 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e4d60def-5dc9-460b-92ed-e3d0157982ec","Type":"ContainerStarted","Data":"715d365ca5f7dd84a91d85b1e876e17deb71037af959f60dad725383a0ccf801"} Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.965965 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-n7gxj" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.965980 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-n7gxj" event={"ID":"538046ba-61c4-43ff-ad1f-a8d80d2ab075","Type":"ContainerDied","Data":"61b1645ad6199f5a1d388077d4b580c92f8d45bff4fda65d7f5262e0a615940a"} Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.966147 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61b1645ad6199f5a1d388077d4b580c92f8d45bff4fda65d7f5262e0a615940a" Feb 20 08:21:48 crc kubenswrapper[4948]: I0220 08:21:48.470441 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-mkpng-config-g4w5d"] Feb 20 08:21:48 crc kubenswrapper[4948]: W0220 08:21:48.479836 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfa672fa0_b98b_477c_a34c_ce996ce6592d.slice/crio-06f713018be060e56a2397e33a80f4b214dc9b77719605af6934b96eaf9c8488 WatchSource:0}: Error finding container 06f713018be060e56a2397e33a80f4b214dc9b77719605af6934b96eaf9c8488: Status 404 returned error can't find the container with id 06f713018be060e56a2397e33a80f4b214dc9b77719605af6934b96eaf9c8488 Feb 20 08:21:48 crc kubenswrapper[4948]: I0220 08:21:48.973400 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-mkpng-config-g4w5d" event={"ID":"fa672fa0-b98b-477c-a34c-ce996ce6592d","Type":"ContainerStarted","Data":"d4c99094574230bbe4d4f5bd6cf1019d664fb9f16355c0c0e8d85cd644e73d32"} Feb 20 08:21:48 crc kubenswrapper[4948]: I0220 08:21:48.973712 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-mkpng-config-g4w5d" event={"ID":"fa672fa0-b98b-477c-a34c-ce996ce6592d","Type":"ContainerStarted","Data":"06f713018be060e56a2397e33a80f4b214dc9b77719605af6934b96eaf9c8488"} Feb 20 08:21:48 crc kubenswrapper[4948]: I0220 08:21:48.976364 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e4d60def-5dc9-460b-92ed-e3d0157982ec","Type":"ContainerStarted","Data":"1c61e52c45386cf02be5fb7e55dbb4aced10eac17718db550b046809d65a5500"} Feb 20 08:21:48 crc kubenswrapper[4948]: I0220 08:21:48.976387 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e4d60def-5dc9-460b-92ed-e3d0157982ec","Type":"ContainerStarted","Data":"10af57300e90334431370f0d766ad2b7c89c11757c0eec9308a6b68829956b58"} Feb 20 08:21:48 crc kubenswrapper[4948]: I0220 08:21:48.976396 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e4d60def-5dc9-460b-92ed-e3d0157982ec","Type":"ContainerStarted","Data":"0ff01f06ef84d4169843f5686becda94c1c31051a153f7883fa0db15d0864b19"} Feb 20 08:21:48 crc kubenswrapper[4948]: I0220 08:21:48.976405 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e4d60def-5dc9-460b-92ed-e3d0157982ec","Type":"ContainerStarted","Data":"ab55b41ba892527d39efc04b2e4d4f30976ca85920ec9c5a5339f15470e49c93"} Feb 20 08:21:48 crc kubenswrapper[4948]: I0220 08:21:48.990381 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-mkpng-config-g4w5d" podStartSLOduration=1.9903651199999999 podStartE2EDuration="1.99036512s" podCreationTimestamp="2026-02-20 08:21:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:21:48.989530439 +0000 UTC m=+957.964025259" watchObservedRunningTime="2026-02-20 08:21:48.99036512 +0000 UTC m=+957.964859940" Feb 20 08:21:49 crc kubenswrapper[4948]: I0220 08:21:49.989985 4948 generic.go:334] "Generic (PLEG): container finished" podID="fa672fa0-b98b-477c-a34c-ce996ce6592d" containerID="d4c99094574230bbe4d4f5bd6cf1019d664fb9f16355c0c0e8d85cd644e73d32" exitCode=0 Feb 20 08:21:49 crc kubenswrapper[4948]: I0220 08:21:49.990037 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-mkpng-config-g4w5d" event={"ID":"fa672fa0-b98b-477c-a34c-ce996ce6592d","Type":"ContainerDied","Data":"d4c99094574230bbe4d4f5bd6cf1019d664fb9f16355c0c0e8d85cd644e73d32"} Feb 20 08:21:50 crc kubenswrapper[4948]: I0220 08:21:50.546638 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-n7gxj"] Feb 20 08:21:50 crc kubenswrapper[4948]: I0220 08:21:50.556057 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-n7gxj"] Feb 20 08:21:51 crc kubenswrapper[4948]: I0220 08:21:51.745165 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="538046ba-61c4-43ff-ad1f-a8d80d2ab075" path="/var/lib/kubelet/pods/538046ba-61c4-43ff-ad1f-a8d80d2ab075/volumes" Feb 20 08:21:52 crc kubenswrapper[4948]: I0220 08:21:52.209811 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-mkpng" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.121276 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.460078 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-97j89"] Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.461604 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-97j89" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.471355 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-97j89"] Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.503164 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.568149 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-0b2f-account-create-update-2f74q"] Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.569396 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-0b2f-account-create-update-2f74q" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.571548 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.585371 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-0b2f-account-create-update-2f74q"] Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.609732 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ced4ac5-e057-476a-908a-f573a8cf997b-operator-scripts\") pod \"cinder-db-create-97j89\" (UID: \"4ced4ac5-e057-476a-908a-f573a8cf997b\") " pod="openstack/cinder-db-create-97j89" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.609917 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrc6c\" (UniqueName: \"kubernetes.io/projected/4ced4ac5-e057-476a-908a-f573a8cf997b-kube-api-access-rrc6c\") pod \"cinder-db-create-97j89\" (UID: \"4ced4ac5-e057-476a-908a-f573a8cf997b\") " pod="openstack/cinder-db-create-97j89" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.644283 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-4djkr"] Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.649345 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4djkr" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.659698 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-4djkr"] Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.698624 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-a86b-account-create-update-mspp7"] Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.699541 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a86b-account-create-update-mspp7" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.701185 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.710799 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6d068f3-210b-41e8-b90a-5406759c0606-operator-scripts\") pod \"cinder-0b2f-account-create-update-2f74q\" (UID: \"b6d068f3-210b-41e8-b90a-5406759c0606\") " pod="openstack/cinder-0b2f-account-create-update-2f74q" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.711003 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrc6c\" (UniqueName: \"kubernetes.io/projected/4ced4ac5-e057-476a-908a-f573a8cf997b-kube-api-access-rrc6c\") pod \"cinder-db-create-97j89\" (UID: \"4ced4ac5-e057-476a-908a-f573a8cf997b\") " pod="openstack/cinder-db-create-97j89" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.711101 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ced4ac5-e057-476a-908a-f573a8cf997b-operator-scripts\") pod \"cinder-db-create-97j89\" (UID: \"4ced4ac5-e057-476a-908a-f573a8cf997b\") " pod="openstack/cinder-db-create-97j89" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.711217 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c86dh\" (UniqueName: \"kubernetes.io/projected/b6d068f3-210b-41e8-b90a-5406759c0606-kube-api-access-c86dh\") pod \"cinder-0b2f-account-create-update-2f74q\" (UID: \"b6d068f3-210b-41e8-b90a-5406759c0606\") " pod="openstack/cinder-0b2f-account-create-update-2f74q" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.712122 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ced4ac5-e057-476a-908a-f573a8cf997b-operator-scripts\") pod \"cinder-db-create-97j89\" (UID: \"4ced4ac5-e057-476a-908a-f573a8cf997b\") " pod="openstack/cinder-db-create-97j89" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.723829 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-a86b-account-create-update-mspp7"] Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.738032 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrc6c\" (UniqueName: \"kubernetes.io/projected/4ced4ac5-e057-476a-908a-f573a8cf997b-kube-api-access-rrc6c\") pod \"cinder-db-create-97j89\" (UID: \"4ced4ac5-e057-476a-908a-f573a8cf997b\") " pod="openstack/cinder-db-create-97j89" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.758219 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-w4bzp"] Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.763478 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-w4bzp" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.772026 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-w4bzp"] Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.785972 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-97j89" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.812759 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10eedd01-3f52-4d44-981d-27bc16af186b-operator-scripts\") pod \"barbican-db-create-4djkr\" (UID: \"10eedd01-3f52-4d44-981d-27bc16af186b\") " pod="openstack/barbican-db-create-4djkr" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.812813 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3f86085-113f-408e-b43c-a499e2fe8b90-operator-scripts\") pod \"barbican-a86b-account-create-update-mspp7\" (UID: \"d3f86085-113f-408e-b43c-a499e2fe8b90\") " pod="openstack/barbican-a86b-account-create-update-mspp7" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.812877 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k984h\" (UniqueName: \"kubernetes.io/projected/d3f86085-113f-408e-b43c-a499e2fe8b90-kube-api-access-k984h\") pod \"barbican-a86b-account-create-update-mspp7\" (UID: \"d3f86085-113f-408e-b43c-a499e2fe8b90\") " pod="openstack/barbican-a86b-account-create-update-mspp7" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.812919 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c86dh\" (UniqueName: \"kubernetes.io/projected/b6d068f3-210b-41e8-b90a-5406759c0606-kube-api-access-c86dh\") pod \"cinder-0b2f-account-create-update-2f74q\" (UID: \"b6d068f3-210b-41e8-b90a-5406759c0606\") " pod="openstack/cinder-0b2f-account-create-update-2f74q" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.812959 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6d068f3-210b-41e8-b90a-5406759c0606-operator-scripts\") pod \"cinder-0b2f-account-create-update-2f74q\" (UID: \"b6d068f3-210b-41e8-b90a-5406759c0606\") " pod="openstack/cinder-0b2f-account-create-update-2f74q" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.813038 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kr2kf\" (UniqueName: \"kubernetes.io/projected/10eedd01-3f52-4d44-981d-27bc16af186b-kube-api-access-kr2kf\") pod \"barbican-db-create-4djkr\" (UID: \"10eedd01-3f52-4d44-981d-27bc16af186b\") " pod="openstack/barbican-db-create-4djkr" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.813956 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6d068f3-210b-41e8-b90a-5406759c0606-operator-scripts\") pod \"cinder-0b2f-account-create-update-2f74q\" (UID: \"b6d068f3-210b-41e8-b90a-5406759c0606\") " pod="openstack/cinder-0b2f-account-create-update-2f74q" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.837164 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-gz5j7"] Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.838201 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c86dh\" (UniqueName: \"kubernetes.io/projected/b6d068f3-210b-41e8-b90a-5406759c0606-kube-api-access-c86dh\") pod \"cinder-0b2f-account-create-update-2f74q\" (UID: \"b6d068f3-210b-41e8-b90a-5406759c0606\") " pod="openstack/cinder-0b2f-account-create-update-2f74q" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.838766 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gz5j7" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.849295 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.849580 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.849685 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.849827 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-v9rj2" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.858941 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-gz5j7"] Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.884546 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-e6d0-account-create-update-cdxd5"] Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.886134 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-e6d0-account-create-update-cdxd5" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.890035 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.895839 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-e6d0-account-create-update-cdxd5"] Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.904806 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-0b2f-account-create-update-2f74q" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.913969 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kr2kf\" (UniqueName: \"kubernetes.io/projected/10eedd01-3f52-4d44-981d-27bc16af186b-kube-api-access-kr2kf\") pod \"barbican-db-create-4djkr\" (UID: \"10eedd01-3f52-4d44-981d-27bc16af186b\") " pod="openstack/barbican-db-create-4djkr" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.914023 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/501e504c-7910-47cb-a181-812a0d72f2db-operator-scripts\") pod \"neutron-db-create-w4bzp\" (UID: \"501e504c-7910-47cb-a181-812a0d72f2db\") " pod="openstack/neutron-db-create-w4bzp" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.914069 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10eedd01-3f52-4d44-981d-27bc16af186b-operator-scripts\") pod \"barbican-db-create-4djkr\" (UID: \"10eedd01-3f52-4d44-981d-27bc16af186b\") " pod="openstack/barbican-db-create-4djkr" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.914091 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3f86085-113f-408e-b43c-a499e2fe8b90-operator-scripts\") pod \"barbican-a86b-account-create-update-mspp7\" (UID: \"d3f86085-113f-408e-b43c-a499e2fe8b90\") " pod="openstack/barbican-a86b-account-create-update-mspp7" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.914242 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k984h\" (UniqueName: \"kubernetes.io/projected/d3f86085-113f-408e-b43c-a499e2fe8b90-kube-api-access-k984h\") pod \"barbican-a86b-account-create-update-mspp7\" (UID: \"d3f86085-113f-408e-b43c-a499e2fe8b90\") " pod="openstack/barbican-a86b-account-create-update-mspp7" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.914311 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvqt6\" (UniqueName: \"kubernetes.io/projected/501e504c-7910-47cb-a181-812a0d72f2db-kube-api-access-bvqt6\") pod \"neutron-db-create-w4bzp\" (UID: \"501e504c-7910-47cb-a181-812a0d72f2db\") " pod="openstack/neutron-db-create-w4bzp" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.914711 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3f86085-113f-408e-b43c-a499e2fe8b90-operator-scripts\") pod \"barbican-a86b-account-create-update-mspp7\" (UID: \"d3f86085-113f-408e-b43c-a499e2fe8b90\") " pod="openstack/barbican-a86b-account-create-update-mspp7" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.914834 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10eedd01-3f52-4d44-981d-27bc16af186b-operator-scripts\") pod \"barbican-db-create-4djkr\" (UID: \"10eedd01-3f52-4d44-981d-27bc16af186b\") " pod="openstack/barbican-db-create-4djkr" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.932008 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kr2kf\" (UniqueName: \"kubernetes.io/projected/10eedd01-3f52-4d44-981d-27bc16af186b-kube-api-access-kr2kf\") pod \"barbican-db-create-4djkr\" (UID: \"10eedd01-3f52-4d44-981d-27bc16af186b\") " pod="openstack/barbican-db-create-4djkr" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.933647 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k984h\" (UniqueName: \"kubernetes.io/projected/d3f86085-113f-408e-b43c-a499e2fe8b90-kube-api-access-k984h\") pod \"barbican-a86b-account-create-update-mspp7\" (UID: \"d3f86085-113f-408e-b43c-a499e2fe8b90\") " pod="openstack/barbican-a86b-account-create-update-mspp7" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.967425 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4djkr" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.016586 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9370de8-8b72-4819-af79-4a9ee1fb6777-combined-ca-bundle\") pod \"keystone-db-sync-gz5j7\" (UID: \"b9370de8-8b72-4819-af79-4a9ee1fb6777\") " pod="openstack/keystone-db-sync-gz5j7" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.016696 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9370de8-8b72-4819-af79-4a9ee1fb6777-config-data\") pod \"keystone-db-sync-gz5j7\" (UID: \"b9370de8-8b72-4819-af79-4a9ee1fb6777\") " pod="openstack/keystone-db-sync-gz5j7" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.016730 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx8w6\" (UniqueName: \"kubernetes.io/projected/b9370de8-8b72-4819-af79-4a9ee1fb6777-kube-api-access-zx8w6\") pod \"keystone-db-sync-gz5j7\" (UID: \"b9370de8-8b72-4819-af79-4a9ee1fb6777\") " pod="openstack/keystone-db-sync-gz5j7" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.016761 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvqt6\" (UniqueName: \"kubernetes.io/projected/501e504c-7910-47cb-a181-812a0d72f2db-kube-api-access-bvqt6\") pod \"neutron-db-create-w4bzp\" (UID: \"501e504c-7910-47cb-a181-812a0d72f2db\") " pod="openstack/neutron-db-create-w4bzp" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.016799 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a739e4d-8960-46f3-8c5d-2373b59165dd-operator-scripts\") pod \"neutron-e6d0-account-create-update-cdxd5\" (UID: \"0a739e4d-8960-46f3-8c5d-2373b59165dd\") " pod="openstack/neutron-e6d0-account-create-update-cdxd5" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.016857 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dqq5\" (UniqueName: \"kubernetes.io/projected/0a739e4d-8960-46f3-8c5d-2373b59165dd-kube-api-access-2dqq5\") pod \"neutron-e6d0-account-create-update-cdxd5\" (UID: \"0a739e4d-8960-46f3-8c5d-2373b59165dd\") " pod="openstack/neutron-e6d0-account-create-update-cdxd5" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.017020 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/501e504c-7910-47cb-a181-812a0d72f2db-operator-scripts\") pod \"neutron-db-create-w4bzp\" (UID: \"501e504c-7910-47cb-a181-812a0d72f2db\") " pod="openstack/neutron-db-create-w4bzp" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.017815 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/501e504c-7910-47cb-a181-812a0d72f2db-operator-scripts\") pod \"neutron-db-create-w4bzp\" (UID: \"501e504c-7910-47cb-a181-812a0d72f2db\") " pod="openstack/neutron-db-create-w4bzp" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.030385 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a86b-account-create-update-mspp7" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.038597 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvqt6\" (UniqueName: \"kubernetes.io/projected/501e504c-7910-47cb-a181-812a0d72f2db-kube-api-access-bvqt6\") pod \"neutron-db-create-w4bzp\" (UID: \"501e504c-7910-47cb-a181-812a0d72f2db\") " pod="openstack/neutron-db-create-w4bzp" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.095438 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-w4bzp" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.118577 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9370de8-8b72-4819-af79-4a9ee1fb6777-combined-ca-bundle\") pod \"keystone-db-sync-gz5j7\" (UID: \"b9370de8-8b72-4819-af79-4a9ee1fb6777\") " pod="openstack/keystone-db-sync-gz5j7" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.118640 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9370de8-8b72-4819-af79-4a9ee1fb6777-config-data\") pod \"keystone-db-sync-gz5j7\" (UID: \"b9370de8-8b72-4819-af79-4a9ee1fb6777\") " pod="openstack/keystone-db-sync-gz5j7" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.118659 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zx8w6\" (UniqueName: \"kubernetes.io/projected/b9370de8-8b72-4819-af79-4a9ee1fb6777-kube-api-access-zx8w6\") pod \"keystone-db-sync-gz5j7\" (UID: \"b9370de8-8b72-4819-af79-4a9ee1fb6777\") " pod="openstack/keystone-db-sync-gz5j7" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.118684 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a739e4d-8960-46f3-8c5d-2373b59165dd-operator-scripts\") pod \"neutron-e6d0-account-create-update-cdxd5\" (UID: \"0a739e4d-8960-46f3-8c5d-2373b59165dd\") " pod="openstack/neutron-e6d0-account-create-update-cdxd5" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.118724 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dqq5\" (UniqueName: \"kubernetes.io/projected/0a739e4d-8960-46f3-8c5d-2373b59165dd-kube-api-access-2dqq5\") pod \"neutron-e6d0-account-create-update-cdxd5\" (UID: \"0a739e4d-8960-46f3-8c5d-2373b59165dd\") " pod="openstack/neutron-e6d0-account-create-update-cdxd5" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.119827 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a739e4d-8960-46f3-8c5d-2373b59165dd-operator-scripts\") pod \"neutron-e6d0-account-create-update-cdxd5\" (UID: \"0a739e4d-8960-46f3-8c5d-2373b59165dd\") " pod="openstack/neutron-e6d0-account-create-update-cdxd5" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.122625 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9370de8-8b72-4819-af79-4a9ee1fb6777-config-data\") pod \"keystone-db-sync-gz5j7\" (UID: \"b9370de8-8b72-4819-af79-4a9ee1fb6777\") " pod="openstack/keystone-db-sync-gz5j7" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.122781 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9370de8-8b72-4819-af79-4a9ee1fb6777-combined-ca-bundle\") pod \"keystone-db-sync-gz5j7\" (UID: \"b9370de8-8b72-4819-af79-4a9ee1fb6777\") " pod="openstack/keystone-db-sync-gz5j7" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.137770 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dqq5\" (UniqueName: \"kubernetes.io/projected/0a739e4d-8960-46f3-8c5d-2373b59165dd-kube-api-access-2dqq5\") pod \"neutron-e6d0-account-create-update-cdxd5\" (UID: \"0a739e4d-8960-46f3-8c5d-2373b59165dd\") " pod="openstack/neutron-e6d0-account-create-update-cdxd5" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.140724 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx8w6\" (UniqueName: \"kubernetes.io/projected/b9370de8-8b72-4819-af79-4a9ee1fb6777-kube-api-access-zx8w6\") pod \"keystone-db-sync-gz5j7\" (UID: \"b9370de8-8b72-4819-af79-4a9ee1fb6777\") " pod="openstack/keystone-db-sync-gz5j7" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.184902 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gz5j7" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.199862 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-e6d0-account-create-update-cdxd5" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.558389 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-zvzvk"] Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.559330 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-zvzvk" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.564400 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.589579 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-zvzvk"] Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.727940 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcv9f\" (UniqueName: \"kubernetes.io/projected/12f1f665-7a04-4794-a727-cc5c060a1933-kube-api-access-rcv9f\") pod \"root-account-create-update-zvzvk\" (UID: \"12f1f665-7a04-4794-a727-cc5c060a1933\") " pod="openstack/root-account-create-update-zvzvk" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.728035 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12f1f665-7a04-4794-a727-cc5c060a1933-operator-scripts\") pod \"root-account-create-update-zvzvk\" (UID: \"12f1f665-7a04-4794-a727-cc5c060a1933\") " pod="openstack/root-account-create-update-zvzvk" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.829870 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcv9f\" (UniqueName: \"kubernetes.io/projected/12f1f665-7a04-4794-a727-cc5c060a1933-kube-api-access-rcv9f\") pod \"root-account-create-update-zvzvk\" (UID: \"12f1f665-7a04-4794-a727-cc5c060a1933\") " pod="openstack/root-account-create-update-zvzvk" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.830163 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12f1f665-7a04-4794-a727-cc5c060a1933-operator-scripts\") pod \"root-account-create-update-zvzvk\" (UID: \"12f1f665-7a04-4794-a727-cc5c060a1933\") " pod="openstack/root-account-create-update-zvzvk" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.830842 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12f1f665-7a04-4794-a727-cc5c060a1933-operator-scripts\") pod \"root-account-create-update-zvzvk\" (UID: \"12f1f665-7a04-4794-a727-cc5c060a1933\") " pod="openstack/root-account-create-update-zvzvk" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.845152 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcv9f\" (UniqueName: \"kubernetes.io/projected/12f1f665-7a04-4794-a727-cc5c060a1933-kube-api-access-rcv9f\") pod \"root-account-create-update-zvzvk\" (UID: \"12f1f665-7a04-4794-a727-cc5c060a1933\") " pod="openstack/root-account-create-update-zvzvk" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.897524 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-zvzvk" Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.547263 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.640159 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8667m\" (UniqueName: \"kubernetes.io/projected/fa672fa0-b98b-477c-a34c-ce996ce6592d-kube-api-access-8667m\") pod \"fa672fa0-b98b-477c-a34c-ce996ce6592d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.641845 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/fa672fa0-b98b-477c-a34c-ce996ce6592d-additional-scripts\") pod \"fa672fa0-b98b-477c-a34c-ce996ce6592d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.641914 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fa672fa0-b98b-477c-a34c-ce996ce6592d-scripts\") pod \"fa672fa0-b98b-477c-a34c-ce996ce6592d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.642044 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-run-ovn\") pod \"fa672fa0-b98b-477c-a34c-ce996ce6592d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.642083 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-log-ovn\") pod \"fa672fa0-b98b-477c-a34c-ce996ce6592d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.642113 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-run\") pod \"fa672fa0-b98b-477c-a34c-ce996ce6592d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.642543 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-run" (OuterVolumeSpecName: "var-run") pod "fa672fa0-b98b-477c-a34c-ce996ce6592d" (UID: "fa672fa0-b98b-477c-a34c-ce996ce6592d"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.643151 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa672fa0-b98b-477c-a34c-ce996ce6592d-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "fa672fa0-b98b-477c-a34c-ce996ce6592d" (UID: "fa672fa0-b98b-477c-a34c-ce996ce6592d"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.643569 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "fa672fa0-b98b-477c-a34c-ce996ce6592d" (UID: "fa672fa0-b98b-477c-a34c-ce996ce6592d"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.643625 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "fa672fa0-b98b-477c-a34c-ce996ce6592d" (UID: "fa672fa0-b98b-477c-a34c-ce996ce6592d"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.644081 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa672fa0-b98b-477c-a34c-ce996ce6592d-scripts" (OuterVolumeSpecName: "scripts") pod "fa672fa0-b98b-477c-a34c-ce996ce6592d" (UID: "fa672fa0-b98b-477c-a34c-ce996ce6592d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.649641 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa672fa0-b98b-477c-a34c-ce996ce6592d-kube-api-access-8667m" (OuterVolumeSpecName: "kube-api-access-8667m") pod "fa672fa0-b98b-477c-a34c-ce996ce6592d" (UID: "fa672fa0-b98b-477c-a34c-ce996ce6592d"). InnerVolumeSpecName "kube-api-access-8667m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.746766 4948 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-log-ovn\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.746798 4948 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-run\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.746809 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8667m\" (UniqueName: \"kubernetes.io/projected/fa672fa0-b98b-477c-a34c-ce996ce6592d-kube-api-access-8667m\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.746821 4948 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/fa672fa0-b98b-477c-a34c-ce996ce6592d-additional-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.746829 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fa672fa0-b98b-477c-a34c-ce996ce6592d-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.746836 4948 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:57 crc kubenswrapper[4948]: I0220 08:21:57.064251 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-mkpng-config-g4w5d" event={"ID":"fa672fa0-b98b-477c-a34c-ce996ce6592d","Type":"ContainerDied","Data":"06f713018be060e56a2397e33a80f4b214dc9b77719605af6934b96eaf9c8488"} Feb 20 08:21:57 crc kubenswrapper[4948]: I0220 08:21:57.064488 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="06f713018be060e56a2397e33a80f4b214dc9b77719605af6934b96eaf9c8488" Feb 20 08:21:57 crc kubenswrapper[4948]: I0220 08:21:57.064547 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:57 crc kubenswrapper[4948]: I0220 08:21:57.148661 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-a86b-account-create-update-mspp7"] Feb 20 08:21:57 crc kubenswrapper[4948]: I0220 08:21:57.382270 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-zvzvk"] Feb 20 08:21:57 crc kubenswrapper[4948]: W0220 08:21:57.385741 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12f1f665_7a04_4794_a727_cc5c060a1933.slice/crio-4ff5b9d3c7487921f2af53697fe4d63849f3213b42b4935658e6dfe3c2d2fd6c WatchSource:0}: Error finding container 4ff5b9d3c7487921f2af53697fe4d63849f3213b42b4935658e6dfe3c2d2fd6c: Status 404 returned error can't find the container with id 4ff5b9d3c7487921f2af53697fe4d63849f3213b42b4935658e6dfe3c2d2fd6c Feb 20 08:21:57 crc kubenswrapper[4948]: I0220 08:21:57.387250 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-e6d0-account-create-update-cdxd5"] Feb 20 08:21:57 crc kubenswrapper[4948]: W0220 08:21:57.391412 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a739e4d_8960_46f3_8c5d_2373b59165dd.slice/crio-1e6ec887d9d3e16f02f5b4d5f7d7d830aca0b26b2e6976000ecd8e9ec7ae4ac5 WatchSource:0}: Error finding container 1e6ec887d9d3e16f02f5b4d5f7d7d830aca0b26b2e6976000ecd8e9ec7ae4ac5: Status 404 returned error can't find the container with id 1e6ec887d9d3e16f02f5b4d5f7d7d830aca0b26b2e6976000ecd8e9ec7ae4ac5 Feb 20 08:21:57 crc kubenswrapper[4948]: I0220 08:21:57.400839 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-4djkr"] Feb 20 08:21:57 crc kubenswrapper[4948]: W0220 08:21:57.424416 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10eedd01_3f52_4d44_981d_27bc16af186b.slice/crio-f4a16b89f1881652510d924497279d30d07a12e1a23c9a2a09460395d54b6754 WatchSource:0}: Error finding container f4a16b89f1881652510d924497279d30d07a12e1a23c9a2a09460395d54b6754: Status 404 returned error can't find the container with id f4a16b89f1881652510d924497279d30d07a12e1a23c9a2a09460395d54b6754 Feb 20 08:21:57 crc kubenswrapper[4948]: I0220 08:21:57.583408 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-97j89"] Feb 20 08:21:57 crc kubenswrapper[4948]: W0220 08:21:57.593083 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ced4ac5_e057_476a_908a_f573a8cf997b.slice/crio-74d816dd0d8f2d728e06f6752ea62cf407348186231cd23345763d2a41ae00ad WatchSource:0}: Error finding container 74d816dd0d8f2d728e06f6752ea62cf407348186231cd23345763d2a41ae00ad: Status 404 returned error can't find the container with id 74d816dd0d8f2d728e06f6752ea62cf407348186231cd23345763d2a41ae00ad Feb 20 08:21:57 crc kubenswrapper[4948]: I0220 08:21:57.597933 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-0b2f-account-create-update-2f74q"] Feb 20 08:21:57 crc kubenswrapper[4948]: W0220 08:21:57.606597 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6d068f3_210b_41e8_b90a_5406759c0606.slice/crio-67031a60eeb343261bf69ab0502142ef436d8c88a1348d9a5bdd6561d7700913 WatchSource:0}: Error finding container 67031a60eeb343261bf69ab0502142ef436d8c88a1348d9a5bdd6561d7700913: Status 404 returned error can't find the container with id 67031a60eeb343261bf69ab0502142ef436d8c88a1348d9a5bdd6561d7700913 Feb 20 08:21:57 crc kubenswrapper[4948]: W0220 08:21:57.612107 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb9370de8_8b72_4819_af79_4a9ee1fb6777.slice/crio-fd31b197071ac286c0663672b75ab91e334d6c66ca4e2bf75a86eb8ab8c0137c WatchSource:0}: Error finding container fd31b197071ac286c0663672b75ab91e334d6c66ca4e2bf75a86eb8ab8c0137c: Status 404 returned error can't find the container with id fd31b197071ac286c0663672b75ab91e334d6c66ca4e2bf75a86eb8ab8c0137c Feb 20 08:21:57 crc kubenswrapper[4948]: I0220 08:21:57.617069 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-w4bzp"] Feb 20 08:21:57 crc kubenswrapper[4948]: I0220 08:21:57.644590 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-gz5j7"] Feb 20 08:21:57 crc kubenswrapper[4948]: I0220 08:21:57.682789 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-mkpng-config-g4w5d"] Feb 20 08:21:57 crc kubenswrapper[4948]: I0220 08:21:57.690001 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-mkpng-config-g4w5d"] Feb 20 08:21:57 crc kubenswrapper[4948]: I0220 08:21:57.799805 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa672fa0-b98b-477c-a34c-ce996ce6592d" path="/var/lib/kubelet/pods/fa672fa0-b98b-477c-a34c-ce996ce6592d/volumes" Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.075516 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-zvzvk" event={"ID":"12f1f665-7a04-4794-a727-cc5c060a1933","Type":"ContainerStarted","Data":"cddaaa0c08e72fa702328c57550853b538c4537573467ac0ebd0761e1fcbc30c"} Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.075781 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-zvzvk" event={"ID":"12f1f665-7a04-4794-a727-cc5c060a1933","Type":"ContainerStarted","Data":"4ff5b9d3c7487921f2af53697fe4d63849f3213b42b4935658e6dfe3c2d2fd6c"} Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.078991 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-pkgzx" event={"ID":"65735728-e734-40de-a185-9caa1db9a47d","Type":"ContainerStarted","Data":"1fb571973e057c8519a5db05840734117d94ed09a1e521d57cf4a92fa88cbcfb"} Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.081666 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-w4bzp" event={"ID":"501e504c-7910-47cb-a181-812a0d72f2db","Type":"ContainerStarted","Data":"f8ab1b0def6ea5b1022acdb9c87659d2c1c20a66a736d1e630b89d097c36d6ef"} Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.081687 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-w4bzp" event={"ID":"501e504c-7910-47cb-a181-812a0d72f2db","Type":"ContainerStarted","Data":"bfa16e47d799d6c141954f2d1749fd776efec59e0c2322be83d41eba7af198db"} Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.090246 4948 generic.go:334] "Generic (PLEG): container finished" podID="10eedd01-3f52-4d44-981d-27bc16af186b" containerID="7e3cc540d00613674fe548789ab2ccc71a06c5714f420cedea97409a8254b14f" exitCode=0 Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.090309 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-4djkr" event={"ID":"10eedd01-3f52-4d44-981d-27bc16af186b","Type":"ContainerDied","Data":"7e3cc540d00613674fe548789ab2ccc71a06c5714f420cedea97409a8254b14f"} Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.090337 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-4djkr" event={"ID":"10eedd01-3f52-4d44-981d-27bc16af186b","Type":"ContainerStarted","Data":"f4a16b89f1881652510d924497279d30d07a12e1a23c9a2a09460395d54b6754"} Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.091556 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gz5j7" event={"ID":"b9370de8-8b72-4819-af79-4a9ee1fb6777","Type":"ContainerStarted","Data":"fd31b197071ac286c0663672b75ab91e334d6c66ca4e2bf75a86eb8ab8c0137c"} Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.093953 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-97j89" event={"ID":"4ced4ac5-e057-476a-908a-f573a8cf997b","Type":"ContainerStarted","Data":"fb257235a0241c571d44f0c21ce5fac86cd6393e8a1250cb5e78d4e1b599fb51"} Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.094042 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-97j89" event={"ID":"4ced4ac5-e057-476a-908a-f573a8cf997b","Type":"ContainerStarted","Data":"74d816dd0d8f2d728e06f6752ea62cf407348186231cd23345763d2a41ae00ad"} Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.102736 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-zvzvk" podStartSLOduration=3.102718112 podStartE2EDuration="3.102718112s" podCreationTimestamp="2026-02-20 08:21:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:21:58.09172735 +0000 UTC m=+967.066222170" watchObservedRunningTime="2026-02-20 08:21:58.102718112 +0000 UTC m=+967.077212932" Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.104118 4948 generic.go:334] "Generic (PLEG): container finished" podID="d3f86085-113f-408e-b43c-a499e2fe8b90" containerID="c565785e94684cef9e257c737e709e95f61b9c29c0e4f1d8e37358207acddb4c" exitCode=0 Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.104185 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-a86b-account-create-update-mspp7" event={"ID":"d3f86085-113f-408e-b43c-a499e2fe8b90","Type":"ContainerDied","Data":"c565785e94684cef9e257c737e709e95f61b9c29c0e4f1d8e37358207acddb4c"} Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.104213 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-a86b-account-create-update-mspp7" event={"ID":"d3f86085-113f-408e-b43c-a499e2fe8b90","Type":"ContainerStarted","Data":"b414d35cd5df6a0028614a71f6eb135810ed8113c4c0d12d90935b0df5cc95f0"} Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.111945 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-0b2f-account-create-update-2f74q" event={"ID":"b6d068f3-210b-41e8-b90a-5406759c0606","Type":"ContainerStarted","Data":"a11f29a4f832ccc8dd04f52fcb05e1f418934455b1143a9d46f2d5ad1bce9425"} Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.112012 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-0b2f-account-create-update-2f74q" event={"ID":"b6d068f3-210b-41e8-b90a-5406759c0606","Type":"ContainerStarted","Data":"67031a60eeb343261bf69ab0502142ef436d8c88a1348d9a5bdd6561d7700913"} Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.117121 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-e6d0-account-create-update-cdxd5" event={"ID":"0a739e4d-8960-46f3-8c5d-2373b59165dd","Type":"ContainerStarted","Data":"ddcd74111702fbb6556b1046c04ec320ea3aecaa381a53da7a6aa1bc0c3d6fd7"} Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.117165 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-e6d0-account-create-update-cdxd5" event={"ID":"0a739e4d-8960-46f3-8c5d-2373b59165dd","Type":"ContainerStarted","Data":"1e6ec887d9d3e16f02f5b4d5f7d7d830aca0b26b2e6976000ecd8e9ec7ae4ac5"} Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.118571 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-w4bzp" podStartSLOduration=4.118551864 podStartE2EDuration="4.118551864s" podCreationTimestamp="2026-02-20 08:21:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:21:58.115212081 +0000 UTC m=+967.089706901" watchObservedRunningTime="2026-02-20 08:21:58.118551864 +0000 UTC m=+967.093046684" Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.137248 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-pkgzx" podStartSLOduration=2.827735116 podStartE2EDuration="16.137230616s" podCreationTimestamp="2026-02-20 08:21:42 +0000 UTC" firstStartedPulling="2026-02-20 08:21:43.400951118 +0000 UTC m=+952.375445938" lastFinishedPulling="2026-02-20 08:21:56.710446618 +0000 UTC m=+965.684941438" observedRunningTime="2026-02-20 08:21:58.134117019 +0000 UTC m=+967.108611839" watchObservedRunningTime="2026-02-20 08:21:58.137230616 +0000 UTC m=+967.111725436" Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.184501 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-0b2f-account-create-update-2f74q" podStartSLOduration=4.184445105 podStartE2EDuration="4.184445105s" podCreationTimestamp="2026-02-20 08:21:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:21:58.183594904 +0000 UTC m=+967.158089724" watchObservedRunningTime="2026-02-20 08:21:58.184445105 +0000 UTC m=+967.158939925" Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.184801 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-e6d0-account-create-update-cdxd5" podStartSLOduration=4.184793683 podStartE2EDuration="4.184793683s" podCreationTimestamp="2026-02-20 08:21:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:21:58.16810899 +0000 UTC m=+967.142603810" watchObservedRunningTime="2026-02-20 08:21:58.184793683 +0000 UTC m=+967.159288493" Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.125098 4948 generic.go:334] "Generic (PLEG): container finished" podID="b6d068f3-210b-41e8-b90a-5406759c0606" containerID="a11f29a4f832ccc8dd04f52fcb05e1f418934455b1143a9d46f2d5ad1bce9425" exitCode=0 Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.125192 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-0b2f-account-create-update-2f74q" event={"ID":"b6d068f3-210b-41e8-b90a-5406759c0606","Type":"ContainerDied","Data":"a11f29a4f832ccc8dd04f52fcb05e1f418934455b1143a9d46f2d5ad1bce9425"} Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.126952 4948 generic.go:334] "Generic (PLEG): container finished" podID="0a739e4d-8960-46f3-8c5d-2373b59165dd" containerID="ddcd74111702fbb6556b1046c04ec320ea3aecaa381a53da7a6aa1bc0c3d6fd7" exitCode=0 Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.127083 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-e6d0-account-create-update-cdxd5" event={"ID":"0a739e4d-8960-46f3-8c5d-2373b59165dd","Type":"ContainerDied","Data":"ddcd74111702fbb6556b1046c04ec320ea3aecaa381a53da7a6aa1bc0c3d6fd7"} Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.128734 4948 generic.go:334] "Generic (PLEG): container finished" podID="501e504c-7910-47cb-a181-812a0d72f2db" containerID="f8ab1b0def6ea5b1022acdb9c87659d2c1c20a66a736d1e630b89d097c36d6ef" exitCode=0 Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.128795 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-w4bzp" event={"ID":"501e504c-7910-47cb-a181-812a0d72f2db","Type":"ContainerDied","Data":"f8ab1b0def6ea5b1022acdb9c87659d2c1c20a66a736d1e630b89d097c36d6ef"} Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.138737 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e4d60def-5dc9-460b-92ed-e3d0157982ec","Type":"ContainerStarted","Data":"0def77fdf4b8be6f4de61d06a4e0076ed0df40fa4e6037fcfcf3199187bc2f02"} Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.138772 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e4d60def-5dc9-460b-92ed-e3d0157982ec","Type":"ContainerStarted","Data":"7be95872e25f6acaeb57ea425561f0eff16c5602eb9dd465855ec55c34f78c37"} Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.138782 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e4d60def-5dc9-460b-92ed-e3d0157982ec","Type":"ContainerStarted","Data":"8b7432bd3983a8cca2ad3cbb32489a07f0374a62db1725e2a564b82efa942d38"} Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.154016 4948 generic.go:334] "Generic (PLEG): container finished" podID="4ced4ac5-e057-476a-908a-f573a8cf997b" containerID="fb257235a0241c571d44f0c21ce5fac86cd6393e8a1250cb5e78d4e1b599fb51" exitCode=0 Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.154061 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-97j89" event={"ID":"4ced4ac5-e057-476a-908a-f573a8cf997b","Type":"ContainerDied","Data":"fb257235a0241c571d44f0c21ce5fac86cd6393e8a1250cb5e78d4e1b599fb51"} Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.155810 4948 generic.go:334] "Generic (PLEG): container finished" podID="12f1f665-7a04-4794-a727-cc5c060a1933" containerID="cddaaa0c08e72fa702328c57550853b538c4537573467ac0ebd0761e1fcbc30c" exitCode=0 Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.155884 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-zvzvk" event={"ID":"12f1f665-7a04-4794-a727-cc5c060a1933","Type":"ContainerDied","Data":"cddaaa0c08e72fa702328c57550853b538c4537573467ac0ebd0761e1fcbc30c"} Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.562193 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a86b-account-create-update-mspp7" Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.662270 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-97j89" Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.668034 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4djkr" Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.716276 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k984h\" (UniqueName: \"kubernetes.io/projected/d3f86085-113f-408e-b43c-a499e2fe8b90-kube-api-access-k984h\") pod \"d3f86085-113f-408e-b43c-a499e2fe8b90\" (UID: \"d3f86085-113f-408e-b43c-a499e2fe8b90\") " Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.716351 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3f86085-113f-408e-b43c-a499e2fe8b90-operator-scripts\") pod \"d3f86085-113f-408e-b43c-a499e2fe8b90\" (UID: \"d3f86085-113f-408e-b43c-a499e2fe8b90\") " Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.717414 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3f86085-113f-408e-b43c-a499e2fe8b90-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d3f86085-113f-408e-b43c-a499e2fe8b90" (UID: "d3f86085-113f-408e-b43c-a499e2fe8b90"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.717906 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3f86085-113f-408e-b43c-a499e2fe8b90-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.721620 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3f86085-113f-408e-b43c-a499e2fe8b90-kube-api-access-k984h" (OuterVolumeSpecName: "kube-api-access-k984h") pod "d3f86085-113f-408e-b43c-a499e2fe8b90" (UID: "d3f86085-113f-408e-b43c-a499e2fe8b90"). InnerVolumeSpecName "kube-api-access-k984h". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.818966 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrc6c\" (UniqueName: \"kubernetes.io/projected/4ced4ac5-e057-476a-908a-f573a8cf997b-kube-api-access-rrc6c\") pod \"4ced4ac5-e057-476a-908a-f573a8cf997b\" (UID: \"4ced4ac5-e057-476a-908a-f573a8cf997b\") " Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.819090 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kr2kf\" (UniqueName: \"kubernetes.io/projected/10eedd01-3f52-4d44-981d-27bc16af186b-kube-api-access-kr2kf\") pod \"10eedd01-3f52-4d44-981d-27bc16af186b\" (UID: \"10eedd01-3f52-4d44-981d-27bc16af186b\") " Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.819133 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ced4ac5-e057-476a-908a-f573a8cf997b-operator-scripts\") pod \"4ced4ac5-e057-476a-908a-f573a8cf997b\" (UID: \"4ced4ac5-e057-476a-908a-f573a8cf997b\") " Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.819319 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10eedd01-3f52-4d44-981d-27bc16af186b-operator-scripts\") pod \"10eedd01-3f52-4d44-981d-27bc16af186b\" (UID: \"10eedd01-3f52-4d44-981d-27bc16af186b\") " Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.819649 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ced4ac5-e057-476a-908a-f573a8cf997b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4ced4ac5-e057-476a-908a-f573a8cf997b" (UID: "4ced4ac5-e057-476a-908a-f573a8cf997b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.820671 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k984h\" (UniqueName: \"kubernetes.io/projected/d3f86085-113f-408e-b43c-a499e2fe8b90-kube-api-access-k984h\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.820720 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ced4ac5-e057-476a-908a-f573a8cf997b-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.821103 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10eedd01-3f52-4d44-981d-27bc16af186b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "10eedd01-3f52-4d44-981d-27bc16af186b" (UID: "10eedd01-3f52-4d44-981d-27bc16af186b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.823267 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10eedd01-3f52-4d44-981d-27bc16af186b-kube-api-access-kr2kf" (OuterVolumeSpecName: "kube-api-access-kr2kf") pod "10eedd01-3f52-4d44-981d-27bc16af186b" (UID: "10eedd01-3f52-4d44-981d-27bc16af186b"). InnerVolumeSpecName "kube-api-access-kr2kf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.824770 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ced4ac5-e057-476a-908a-f573a8cf997b-kube-api-access-rrc6c" (OuterVolumeSpecName: "kube-api-access-rrc6c") pod "4ced4ac5-e057-476a-908a-f573a8cf997b" (UID: "4ced4ac5-e057-476a-908a-f573a8cf997b"). InnerVolumeSpecName "kube-api-access-rrc6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.922951 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10eedd01-3f52-4d44-981d-27bc16af186b-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.923018 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrc6c\" (UniqueName: \"kubernetes.io/projected/4ced4ac5-e057-476a-908a-f573a8cf997b-kube-api-access-rrc6c\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.923035 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kr2kf\" (UniqueName: \"kubernetes.io/projected/10eedd01-3f52-4d44-981d-27bc16af186b-kube-api-access-kr2kf\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.165014 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-97j89" event={"ID":"4ced4ac5-e057-476a-908a-f573a8cf997b","Type":"ContainerDied","Data":"74d816dd0d8f2d728e06f6752ea62cf407348186231cd23345763d2a41ae00ad"} Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.166220 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74d816dd0d8f2d728e06f6752ea62cf407348186231cd23345763d2a41ae00ad" Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.166425 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-97j89" Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.192381 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-a86b-account-create-update-mspp7" event={"ID":"d3f86085-113f-408e-b43c-a499e2fe8b90","Type":"ContainerDied","Data":"b414d35cd5df6a0028614a71f6eb135810ed8113c4c0d12d90935b0df5cc95f0"} Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.192443 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b414d35cd5df6a0028614a71f6eb135810ed8113c4c0d12d90935b0df5cc95f0" Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.192405 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a86b-account-create-update-mspp7" Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.199775 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-4djkr" event={"ID":"10eedd01-3f52-4d44-981d-27bc16af186b","Type":"ContainerDied","Data":"f4a16b89f1881652510d924497279d30d07a12e1a23c9a2a09460395d54b6754"} Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.199826 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4a16b89f1881652510d924497279d30d07a12e1a23c9a2a09460395d54b6754" Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.200033 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4djkr" Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.205354 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e4d60def-5dc9-460b-92ed-e3d0157982ec","Type":"ContainerStarted","Data":"f0e914fe99d8fbabaf9e4af16ab7bd5f4b91015e66934992a5c19a45d77421d7"} Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.634165 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-e6d0-account-create-update-cdxd5" Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.741183 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dqq5\" (UniqueName: \"kubernetes.io/projected/0a739e4d-8960-46f3-8c5d-2373b59165dd-kube-api-access-2dqq5\") pod \"0a739e4d-8960-46f3-8c5d-2373b59165dd\" (UID: \"0a739e4d-8960-46f3-8c5d-2373b59165dd\") " Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.741461 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a739e4d-8960-46f3-8c5d-2373b59165dd-operator-scripts\") pod \"0a739e4d-8960-46f3-8c5d-2373b59165dd\" (UID: \"0a739e4d-8960-46f3-8c5d-2373b59165dd\") " Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.742899 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a739e4d-8960-46f3-8c5d-2373b59165dd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0a739e4d-8960-46f3-8c5d-2373b59165dd" (UID: "0a739e4d-8960-46f3-8c5d-2373b59165dd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.766236 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a739e4d-8960-46f3-8c5d-2373b59165dd-kube-api-access-2dqq5" (OuterVolumeSpecName: "kube-api-access-2dqq5") pod "0a739e4d-8960-46f3-8c5d-2373b59165dd" (UID: "0a739e4d-8960-46f3-8c5d-2373b59165dd"). InnerVolumeSpecName "kube-api-access-2dqq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.843453 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dqq5\" (UniqueName: \"kubernetes.io/projected/0a739e4d-8960-46f3-8c5d-2373b59165dd-kube-api-access-2dqq5\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.843486 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a739e4d-8960-46f3-8c5d-2373b59165dd-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.886402 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-w4bzp" Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.894892 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-0b2f-account-create-update-2f74q" Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.900158 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-zvzvk" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.046596 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6d068f3-210b-41e8-b90a-5406759c0606-operator-scripts\") pod \"b6d068f3-210b-41e8-b90a-5406759c0606\" (UID: \"b6d068f3-210b-41e8-b90a-5406759c0606\") " Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.046685 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12f1f665-7a04-4794-a727-cc5c060a1933-operator-scripts\") pod \"12f1f665-7a04-4794-a727-cc5c060a1933\" (UID: \"12f1f665-7a04-4794-a727-cc5c060a1933\") " Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.046726 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/501e504c-7910-47cb-a181-812a0d72f2db-operator-scripts\") pod \"501e504c-7910-47cb-a181-812a0d72f2db\" (UID: \"501e504c-7910-47cb-a181-812a0d72f2db\") " Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.046793 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcv9f\" (UniqueName: \"kubernetes.io/projected/12f1f665-7a04-4794-a727-cc5c060a1933-kube-api-access-rcv9f\") pod \"12f1f665-7a04-4794-a727-cc5c060a1933\" (UID: \"12f1f665-7a04-4794-a727-cc5c060a1933\") " Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.046860 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvqt6\" (UniqueName: \"kubernetes.io/projected/501e504c-7910-47cb-a181-812a0d72f2db-kube-api-access-bvqt6\") pod \"501e504c-7910-47cb-a181-812a0d72f2db\" (UID: \"501e504c-7910-47cb-a181-812a0d72f2db\") " Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.046883 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c86dh\" (UniqueName: \"kubernetes.io/projected/b6d068f3-210b-41e8-b90a-5406759c0606-kube-api-access-c86dh\") pod \"b6d068f3-210b-41e8-b90a-5406759c0606\" (UID: \"b6d068f3-210b-41e8-b90a-5406759c0606\") " Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.047611 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/501e504c-7910-47cb-a181-812a0d72f2db-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "501e504c-7910-47cb-a181-812a0d72f2db" (UID: "501e504c-7910-47cb-a181-812a0d72f2db"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.047696 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6d068f3-210b-41e8-b90a-5406759c0606-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b6d068f3-210b-41e8-b90a-5406759c0606" (UID: "b6d068f3-210b-41e8-b90a-5406759c0606"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.047917 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12f1f665-7a04-4794-a727-cc5c060a1933-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "12f1f665-7a04-4794-a727-cc5c060a1933" (UID: "12f1f665-7a04-4794-a727-cc5c060a1933"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.053202 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6d068f3-210b-41e8-b90a-5406759c0606-kube-api-access-c86dh" (OuterVolumeSpecName: "kube-api-access-c86dh") pod "b6d068f3-210b-41e8-b90a-5406759c0606" (UID: "b6d068f3-210b-41e8-b90a-5406759c0606"). InnerVolumeSpecName "kube-api-access-c86dh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.053295 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12f1f665-7a04-4794-a727-cc5c060a1933-kube-api-access-rcv9f" (OuterVolumeSpecName: "kube-api-access-rcv9f") pod "12f1f665-7a04-4794-a727-cc5c060a1933" (UID: "12f1f665-7a04-4794-a727-cc5c060a1933"). InnerVolumeSpecName "kube-api-access-rcv9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.058950 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/501e504c-7910-47cb-a181-812a0d72f2db-kube-api-access-bvqt6" (OuterVolumeSpecName: "kube-api-access-bvqt6") pod "501e504c-7910-47cb-a181-812a0d72f2db" (UID: "501e504c-7910-47cb-a181-812a0d72f2db"). InnerVolumeSpecName "kube-api-access-bvqt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.149106 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12f1f665-7a04-4794-a727-cc5c060a1933-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.149142 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/501e504c-7910-47cb-a181-812a0d72f2db-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.149151 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcv9f\" (UniqueName: \"kubernetes.io/projected/12f1f665-7a04-4794-a727-cc5c060a1933-kube-api-access-rcv9f\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.149163 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvqt6\" (UniqueName: \"kubernetes.io/projected/501e504c-7910-47cb-a181-812a0d72f2db-kube-api-access-bvqt6\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.149172 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c86dh\" (UniqueName: \"kubernetes.io/projected/b6d068f3-210b-41e8-b90a-5406759c0606-kube-api-access-c86dh\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.149180 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6d068f3-210b-41e8-b90a-5406759c0606-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.228186 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e4d60def-5dc9-460b-92ed-e3d0157982ec","Type":"ContainerStarted","Data":"3355e23b27ec2e8d34fef5fb30fe085eb1135b951828ca9cccf056bb4392c599"} Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.228230 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e4d60def-5dc9-460b-92ed-e3d0157982ec","Type":"ContainerStarted","Data":"73d89eec04dd76cb9501192d2204144dca67ce3fa8d711ecff913828c149145f"} Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.233945 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-zvzvk" event={"ID":"12f1f665-7a04-4794-a727-cc5c060a1933","Type":"ContainerDied","Data":"4ff5b9d3c7487921f2af53697fe4d63849f3213b42b4935658e6dfe3c2d2fd6c"} Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.234033 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ff5b9d3c7487921f2af53697fe4d63849f3213b42b4935658e6dfe3c2d2fd6c" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.233958 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-zvzvk" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.236261 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-e6d0-account-create-update-cdxd5" event={"ID":"0a739e4d-8960-46f3-8c5d-2373b59165dd","Type":"ContainerDied","Data":"1e6ec887d9d3e16f02f5b4d5f7d7d830aca0b26b2e6976000ecd8e9ec7ae4ac5"} Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.236294 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e6ec887d9d3e16f02f5b4d5f7d7d830aca0b26b2e6976000ecd8e9ec7ae4ac5" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.236360 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-e6d0-account-create-update-cdxd5" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.241933 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-0b2f-account-create-update-2f74q" event={"ID":"b6d068f3-210b-41e8-b90a-5406759c0606","Type":"ContainerDied","Data":"67031a60eeb343261bf69ab0502142ef436d8c88a1348d9a5bdd6561d7700913"} Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.241991 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67031a60eeb343261bf69ab0502142ef436d8c88a1348d9a5bdd6561d7700913" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.242046 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-0b2f-account-create-update-2f74q" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.247286 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-w4bzp" event={"ID":"501e504c-7910-47cb-a181-812a0d72f2db","Type":"ContainerDied","Data":"bfa16e47d799d6c141954f2d1749fd776efec59e0c2322be83d41eba7af198db"} Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.247321 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bfa16e47d799d6c141954f2d1749fd776efec59e0c2322be83d41eba7af198db" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.247370 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-w4bzp" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.444547 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lnp7g"] Feb 20 08:22:03 crc kubenswrapper[4948]: E0220 08:22:03.445311 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12f1f665-7a04-4794-a727-cc5c060a1933" containerName="mariadb-account-create-update" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.445329 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="12f1f665-7a04-4794-a727-cc5c060a1933" containerName="mariadb-account-create-update" Feb 20 08:22:03 crc kubenswrapper[4948]: E0220 08:22:03.445347 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ced4ac5-e057-476a-908a-f573a8cf997b" containerName="mariadb-database-create" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.445357 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ced4ac5-e057-476a-908a-f573a8cf997b" containerName="mariadb-database-create" Feb 20 08:22:03 crc kubenswrapper[4948]: E0220 08:22:03.445370 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="501e504c-7910-47cb-a181-812a0d72f2db" containerName="mariadb-database-create" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.445378 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="501e504c-7910-47cb-a181-812a0d72f2db" containerName="mariadb-database-create" Feb 20 08:22:03 crc kubenswrapper[4948]: E0220 08:22:03.445396 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6d068f3-210b-41e8-b90a-5406759c0606" containerName="mariadb-account-create-update" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.445403 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6d068f3-210b-41e8-b90a-5406759c0606" containerName="mariadb-account-create-update" Feb 20 08:22:03 crc kubenswrapper[4948]: E0220 08:22:03.445419 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10eedd01-3f52-4d44-981d-27bc16af186b" containerName="mariadb-database-create" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.445427 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="10eedd01-3f52-4d44-981d-27bc16af186b" containerName="mariadb-database-create" Feb 20 08:22:03 crc kubenswrapper[4948]: E0220 08:22:03.445446 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3f86085-113f-408e-b43c-a499e2fe8b90" containerName="mariadb-account-create-update" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.445455 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3f86085-113f-408e-b43c-a499e2fe8b90" containerName="mariadb-account-create-update" Feb 20 08:22:03 crc kubenswrapper[4948]: E0220 08:22:03.445464 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa672fa0-b98b-477c-a34c-ce996ce6592d" containerName="ovn-config" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.445472 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa672fa0-b98b-477c-a34c-ce996ce6592d" containerName="ovn-config" Feb 20 08:22:03 crc kubenswrapper[4948]: E0220 08:22:03.445491 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a739e4d-8960-46f3-8c5d-2373b59165dd" containerName="mariadb-account-create-update" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.445499 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a739e4d-8960-46f3-8c5d-2373b59165dd" containerName="mariadb-account-create-update" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.445729 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="10eedd01-3f52-4d44-981d-27bc16af186b" containerName="mariadb-database-create" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.445747 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a739e4d-8960-46f3-8c5d-2373b59165dd" containerName="mariadb-account-create-update" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.445754 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="501e504c-7910-47cb-a181-812a0d72f2db" containerName="mariadb-database-create" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.445765 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6d068f3-210b-41e8-b90a-5406759c0606" containerName="mariadb-account-create-update" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.445782 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa672fa0-b98b-477c-a34c-ce996ce6592d" containerName="ovn-config" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.445791 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3f86085-113f-408e-b43c-a499e2fe8b90" containerName="mariadb-account-create-update" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.445802 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="12f1f665-7a04-4794-a727-cc5c060a1933" containerName="mariadb-account-create-update" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.445811 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ced4ac5-e057-476a-908a-f573a8cf997b" containerName="mariadb-database-create" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.454474 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lnp7g" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.495055 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lnp7g"] Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.585080 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81e7d5de-2379-46c7-ba9b-ffcacad79e29-utilities\") pod \"redhat-marketplace-lnp7g\" (UID: \"81e7d5de-2379-46c7-ba9b-ffcacad79e29\") " pod="openshift-marketplace/redhat-marketplace-lnp7g" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.585119 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81e7d5de-2379-46c7-ba9b-ffcacad79e29-catalog-content\") pod \"redhat-marketplace-lnp7g\" (UID: \"81e7d5de-2379-46c7-ba9b-ffcacad79e29\") " pod="openshift-marketplace/redhat-marketplace-lnp7g" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.585144 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5vn9\" (UniqueName: \"kubernetes.io/projected/81e7d5de-2379-46c7-ba9b-ffcacad79e29-kube-api-access-w5vn9\") pod \"redhat-marketplace-lnp7g\" (UID: \"81e7d5de-2379-46c7-ba9b-ffcacad79e29\") " pod="openshift-marketplace/redhat-marketplace-lnp7g" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.686906 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81e7d5de-2379-46c7-ba9b-ffcacad79e29-utilities\") pod \"redhat-marketplace-lnp7g\" (UID: \"81e7d5de-2379-46c7-ba9b-ffcacad79e29\") " pod="openshift-marketplace/redhat-marketplace-lnp7g" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.686938 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81e7d5de-2379-46c7-ba9b-ffcacad79e29-catalog-content\") pod \"redhat-marketplace-lnp7g\" (UID: \"81e7d5de-2379-46c7-ba9b-ffcacad79e29\") " pod="openshift-marketplace/redhat-marketplace-lnp7g" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.686959 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5vn9\" (UniqueName: \"kubernetes.io/projected/81e7d5de-2379-46c7-ba9b-ffcacad79e29-kube-api-access-w5vn9\") pod \"redhat-marketplace-lnp7g\" (UID: \"81e7d5de-2379-46c7-ba9b-ffcacad79e29\") " pod="openshift-marketplace/redhat-marketplace-lnp7g" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.687740 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81e7d5de-2379-46c7-ba9b-ffcacad79e29-catalog-content\") pod \"redhat-marketplace-lnp7g\" (UID: \"81e7d5de-2379-46c7-ba9b-ffcacad79e29\") " pod="openshift-marketplace/redhat-marketplace-lnp7g" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.687835 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81e7d5de-2379-46c7-ba9b-ffcacad79e29-utilities\") pod \"redhat-marketplace-lnp7g\" (UID: \"81e7d5de-2379-46c7-ba9b-ffcacad79e29\") " pod="openshift-marketplace/redhat-marketplace-lnp7g" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.716533 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5vn9\" (UniqueName: \"kubernetes.io/projected/81e7d5de-2379-46c7-ba9b-ffcacad79e29-kube-api-access-w5vn9\") pod \"redhat-marketplace-lnp7g\" (UID: \"81e7d5de-2379-46c7-ba9b-ffcacad79e29\") " pod="openshift-marketplace/redhat-marketplace-lnp7g" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.796416 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lnp7g" Feb 20 08:22:04 crc kubenswrapper[4948]: I0220 08:22:04.256875 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lnp7g"] Feb 20 08:22:04 crc kubenswrapper[4948]: W0220 08:22:04.266236 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod81e7d5de_2379_46c7_ba9b_ffcacad79e29.slice/crio-ac4ceb2acdc56fbdc4b52d619a840394196af858618bbcf55bc7c1dc060d1a2f WatchSource:0}: Error finding container ac4ceb2acdc56fbdc4b52d619a840394196af858618bbcf55bc7c1dc060d1a2f: Status 404 returned error can't find the container with id ac4ceb2acdc56fbdc4b52d619a840394196af858618bbcf55bc7c1dc060d1a2f Feb 20 08:22:04 crc kubenswrapper[4948]: I0220 08:22:04.278565 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e4d60def-5dc9-460b-92ed-e3d0157982ec","Type":"ContainerStarted","Data":"5369372614d8e43350703b61475cc193540f62c9e2a5e484b1fbcd3c24f8cf51"} Feb 20 08:22:04 crc kubenswrapper[4948]: I0220 08:22:04.278944 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e4d60def-5dc9-460b-92ed-e3d0157982ec","Type":"ContainerStarted","Data":"6c3625774a56cd70321fcb0287393d4cc54f0526941e165f455cd535741f10cd"} Feb 20 08:22:04 crc kubenswrapper[4948]: I0220 08:22:04.280624 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gz5j7" event={"ID":"b9370de8-8b72-4819-af79-4a9ee1fb6777","Type":"ContainerStarted","Data":"72b50822f4c48e3166b0066730c258d864d6094e98d9d9617ca0abdc1f36ae38"} Feb 20 08:22:04 crc kubenswrapper[4948]: I0220 08:22:04.307272 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-gz5j7" podStartSLOduration=4.165632658 podStartE2EDuration="10.307252965s" podCreationTimestamp="2026-02-20 08:21:54 +0000 UTC" firstStartedPulling="2026-02-20 08:21:57.624738803 +0000 UTC m=+966.599233623" lastFinishedPulling="2026-02-20 08:22:03.76635911 +0000 UTC m=+972.740853930" observedRunningTime="2026-02-20 08:22:04.303165644 +0000 UTC m=+973.277660464" watchObservedRunningTime="2026-02-20 08:22:04.307252965 +0000 UTC m=+973.281747785" Feb 20 08:22:05 crc kubenswrapper[4948]: I0220 08:22:05.299865 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e4d60def-5dc9-460b-92ed-e3d0157982ec","Type":"ContainerStarted","Data":"f5d776784274a96fef8898bd072052c1535eba851190d0eacd8ac02fe6ab20f6"} Feb 20 08:22:05 crc kubenswrapper[4948]: I0220 08:22:05.300017 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e4d60def-5dc9-460b-92ed-e3d0157982ec","Type":"ContainerStarted","Data":"e92a5dd37239b7f9de68a6175ce427a7d03ddfec618577a2b42564193f86893a"} Feb 20 08:22:05 crc kubenswrapper[4948]: I0220 08:22:05.301880 4948 generic.go:334] "Generic (PLEG): container finished" podID="81e7d5de-2379-46c7-ba9b-ffcacad79e29" containerID="82db4fd040d5ae64aa60b6299d284256aa892181e4cf858ebfad62dc9681817b" exitCode=0 Feb 20 08:22:05 crc kubenswrapper[4948]: I0220 08:22:05.302067 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnp7g" event={"ID":"81e7d5de-2379-46c7-ba9b-ffcacad79e29","Type":"ContainerDied","Data":"82db4fd040d5ae64aa60b6299d284256aa892181e4cf858ebfad62dc9681817b"} Feb 20 08:22:05 crc kubenswrapper[4948]: I0220 08:22:05.302096 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnp7g" event={"ID":"81e7d5de-2379-46c7-ba9b-ffcacad79e29","Type":"ContainerStarted","Data":"ac4ceb2acdc56fbdc4b52d619a840394196af858618bbcf55bc7c1dc060d1a2f"} Feb 20 08:22:09 crc kubenswrapper[4948]: I0220 08:22:09.340378 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e4d60def-5dc9-460b-92ed-e3d0157982ec","Type":"ContainerStarted","Data":"38118f7052314e9310bad19f8bc505ccd82a8634fb080d44e8b173fffa583cd6"} Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.349420 4948 generic.go:334] "Generic (PLEG): container finished" podID="81e7d5de-2379-46c7-ba9b-ffcacad79e29" containerID="833390a5ef5c4525790a7cad88740d3ce4d0df639a0829042164aca3b32d8d15" exitCode=0 Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.349469 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnp7g" event={"ID":"81e7d5de-2379-46c7-ba9b-ffcacad79e29","Type":"ContainerDied","Data":"833390a5ef5c4525790a7cad88740d3ce4d0df639a0829042164aca3b32d8d15"} Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.352601 4948 generic.go:334] "Generic (PLEG): container finished" podID="65735728-e734-40de-a185-9caa1db9a47d" containerID="1fb571973e057c8519a5db05840734117d94ed09a1e521d57cf4a92fa88cbcfb" exitCode=0 Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.352687 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-pkgzx" event={"ID":"65735728-e734-40de-a185-9caa1db9a47d","Type":"ContainerDied","Data":"1fb571973e057c8519a5db05840734117d94ed09a1e521d57cf4a92fa88cbcfb"} Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.405108 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=27.962875945 podStartE2EDuration="41.405082558s" podCreationTimestamp="2026-02-20 08:21:29 +0000 UTC" firstStartedPulling="2026-02-20 08:21:47.117874941 +0000 UTC m=+956.092369761" lastFinishedPulling="2026-02-20 08:22:00.560081544 +0000 UTC m=+969.534576374" observedRunningTime="2026-02-20 08:22:10.399139931 +0000 UTC m=+979.373634761" watchObservedRunningTime="2026-02-20 08:22:10.405082558 +0000 UTC m=+979.379577398" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.698203 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-mvrcc"] Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.699760 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.708181 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.716267 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-mvrcc"] Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.798987 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-dns-svc\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.799340 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.799405 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-config\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.799436 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.799497 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.799565 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57xf5\" (UniqueName: \"kubernetes.io/projected/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-kube-api-access-57xf5\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.900546 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.900607 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57xf5\" (UniqueName: \"kubernetes.io/projected/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-kube-api-access-57xf5\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.900652 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-dns-svc\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.900668 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.900692 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-config\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.900718 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.901619 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.901624 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-dns-svc\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.901775 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.901827 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-config\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.903494 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.922361 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57xf5\" (UniqueName: \"kubernetes.io/projected/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-kube-api-access-57xf5\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.024230 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.362144 4948 generic.go:334] "Generic (PLEG): container finished" podID="b9370de8-8b72-4819-af79-4a9ee1fb6777" containerID="72b50822f4c48e3166b0066730c258d864d6094e98d9d9617ca0abdc1f36ae38" exitCode=0 Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.362222 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gz5j7" event={"ID":"b9370de8-8b72-4819-af79-4a9ee1fb6777","Type":"ContainerDied","Data":"72b50822f4c48e3166b0066730c258d864d6094e98d9d9617ca0abdc1f36ae38"} Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.364563 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnp7g" event={"ID":"81e7d5de-2379-46c7-ba9b-ffcacad79e29","Type":"ContainerStarted","Data":"59eceb81df286be225188d13264d44596719c0a06ca1c39903c517b422013a68"} Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.402195 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lnp7g" podStartSLOduration=2.957574144 podStartE2EDuration="8.402179142s" podCreationTimestamp="2026-02-20 08:22:03 +0000 UTC" firstStartedPulling="2026-02-20 08:22:05.303570081 +0000 UTC m=+974.278064901" lastFinishedPulling="2026-02-20 08:22:10.748175079 +0000 UTC m=+979.722669899" observedRunningTime="2026-02-20 08:22:11.39563659 +0000 UTC m=+980.370131410" watchObservedRunningTime="2026-02-20 08:22:11.402179142 +0000 UTC m=+980.376673962" Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.548678 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-mvrcc"] Feb 20 08:22:11 crc kubenswrapper[4948]: W0220 08:22:11.551211 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b8788ed_d6f3_4344_b8b2_7bb554ecd25e.slice/crio-5b5cca04b9c3f166250baa067797237d67a3b5df1ae92815a7da886a955a6658 WatchSource:0}: Error finding container 5b5cca04b9c3f166250baa067797237d67a3b5df1ae92815a7da886a955a6658: Status 404 returned error can't find the container with id 5b5cca04b9c3f166250baa067797237d67a3b5df1ae92815a7da886a955a6658 Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.714729 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-pkgzx" Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.828440 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-config-data\") pod \"65735728-e734-40de-a185-9caa1db9a47d\" (UID: \"65735728-e734-40de-a185-9caa1db9a47d\") " Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.828578 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j589d\" (UniqueName: \"kubernetes.io/projected/65735728-e734-40de-a185-9caa1db9a47d-kube-api-access-j589d\") pod \"65735728-e734-40de-a185-9caa1db9a47d\" (UID: \"65735728-e734-40de-a185-9caa1db9a47d\") " Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.828602 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-db-sync-config-data\") pod \"65735728-e734-40de-a185-9caa1db9a47d\" (UID: \"65735728-e734-40de-a185-9caa1db9a47d\") " Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.828681 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-combined-ca-bundle\") pod \"65735728-e734-40de-a185-9caa1db9a47d\" (UID: \"65735728-e734-40de-a185-9caa1db9a47d\") " Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.833647 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "65735728-e734-40de-a185-9caa1db9a47d" (UID: "65735728-e734-40de-a185-9caa1db9a47d"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.834007 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65735728-e734-40de-a185-9caa1db9a47d-kube-api-access-j589d" (OuterVolumeSpecName: "kube-api-access-j589d") pod "65735728-e734-40de-a185-9caa1db9a47d" (UID: "65735728-e734-40de-a185-9caa1db9a47d"). InnerVolumeSpecName "kube-api-access-j589d". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.850375 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "65735728-e734-40de-a185-9caa1db9a47d" (UID: "65735728-e734-40de-a185-9caa1db9a47d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.869861 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-config-data" (OuterVolumeSpecName: "config-data") pod "65735728-e734-40de-a185-9caa1db9a47d" (UID: "65735728-e734-40de-a185-9caa1db9a47d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.931280 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j589d\" (UniqueName: \"kubernetes.io/projected/65735728-e734-40de-a185-9caa1db9a47d-kube-api-access-j589d\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.931336 4948 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.931357 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.931374 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.375771 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-pkgzx" event={"ID":"65735728-e734-40de-a185-9caa1db9a47d","Type":"ContainerDied","Data":"5e99cbf3924927a30ad54b19ba1a0b9e160811b4e81053363ade5b0662fa69fa"} Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.375821 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e99cbf3924927a30ad54b19ba1a0b9e160811b4e81053363ade5b0662fa69fa" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.375912 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-pkgzx" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.378705 4948 generic.go:334] "Generic (PLEG): container finished" podID="0b8788ed-d6f3-4344-b8b2-7bb554ecd25e" containerID="596bfc5f705d539b55d6de0f0ebc270b6e2a7ef5c99c535a13f5bc566d754f58" exitCode=0 Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.378753 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" event={"ID":"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e","Type":"ContainerDied","Data":"596bfc5f705d539b55d6de0f0ebc270b6e2a7ef5c99c535a13f5bc566d754f58"} Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.378921 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" event={"ID":"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e","Type":"ContainerStarted","Data":"5b5cca04b9c3f166250baa067797237d67a3b5df1ae92815a7da886a955a6658"} Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.763878 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-mvrcc"] Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.769891 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gz5j7" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.811455 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-v6srk"] Feb 20 08:22:12 crc kubenswrapper[4948]: E0220 08:22:12.811927 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65735728-e734-40de-a185-9caa1db9a47d" containerName="glance-db-sync" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.811954 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="65735728-e734-40de-a185-9caa1db9a47d" containerName="glance-db-sync" Feb 20 08:22:12 crc kubenswrapper[4948]: E0220 08:22:12.811967 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9370de8-8b72-4819-af79-4a9ee1fb6777" containerName="keystone-db-sync" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.812027 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9370de8-8b72-4819-af79-4a9ee1fb6777" containerName="keystone-db-sync" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.812231 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="65735728-e734-40de-a185-9caa1db9a47d" containerName="glance-db-sync" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.812264 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9370de8-8b72-4819-af79-4a9ee1fb6777" containerName="keystone-db-sync" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.813334 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.822033 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-v6srk"] Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.847724 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9370de8-8b72-4819-af79-4a9ee1fb6777-config-data\") pod \"b9370de8-8b72-4819-af79-4a9ee1fb6777\" (UID: \"b9370de8-8b72-4819-af79-4a9ee1fb6777\") " Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.847866 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zx8w6\" (UniqueName: \"kubernetes.io/projected/b9370de8-8b72-4819-af79-4a9ee1fb6777-kube-api-access-zx8w6\") pod \"b9370de8-8b72-4819-af79-4a9ee1fb6777\" (UID: \"b9370de8-8b72-4819-af79-4a9ee1fb6777\") " Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.847921 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9370de8-8b72-4819-af79-4a9ee1fb6777-combined-ca-bundle\") pod \"b9370de8-8b72-4819-af79-4a9ee1fb6777\" (UID: \"b9370de8-8b72-4819-af79-4a9ee1fb6777\") " Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.852693 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9370de8-8b72-4819-af79-4a9ee1fb6777-kube-api-access-zx8w6" (OuterVolumeSpecName: "kube-api-access-zx8w6") pod "b9370de8-8b72-4819-af79-4a9ee1fb6777" (UID: "b9370de8-8b72-4819-af79-4a9ee1fb6777"). InnerVolumeSpecName "kube-api-access-zx8w6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.875346 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9370de8-8b72-4819-af79-4a9ee1fb6777-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b9370de8-8b72-4819-af79-4a9ee1fb6777" (UID: "b9370de8-8b72-4819-af79-4a9ee1fb6777"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.891446 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9370de8-8b72-4819-af79-4a9ee1fb6777-config-data" (OuterVolumeSpecName: "config-data") pod "b9370de8-8b72-4819-af79-4a9ee1fb6777" (UID: "b9370de8-8b72-4819-af79-4a9ee1fb6777"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.950408 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.950501 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.950651 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74rwq\" (UniqueName: \"kubernetes.io/projected/7960b704-21ac-4d75-8226-a9bbae4c2b50-kube-api-access-74rwq\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.950729 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-config\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.950833 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.950887 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.951089 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9370de8-8b72-4819-af79-4a9ee1fb6777-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.951101 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9370de8-8b72-4819-af79-4a9ee1fb6777-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.951111 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zx8w6\" (UniqueName: \"kubernetes.io/projected/b9370de8-8b72-4819-af79-4a9ee1fb6777-kube-api-access-zx8w6\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.052360 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.052415 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.052494 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.052516 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.052553 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74rwq\" (UniqueName: \"kubernetes.io/projected/7960b704-21ac-4d75-8226-a9bbae4c2b50-kube-api-access-74rwq\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.052590 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-config\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.053512 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.054032 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.054049 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.054098 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-config\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.054360 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.085902 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74rwq\" (UniqueName: \"kubernetes.io/projected/7960b704-21ac-4d75-8226-a9bbae4c2b50-kube-api-access-74rwq\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.232120 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.387931 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gz5j7" event={"ID":"b9370de8-8b72-4819-af79-4a9ee1fb6777","Type":"ContainerDied","Data":"fd31b197071ac286c0663672b75ab91e334d6c66ca4e2bf75a86eb8ab8c0137c"} Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.388013 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gz5j7" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.388032 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd31b197071ac286c0663672b75ab91e334d6c66ca4e2bf75a86eb8ab8c0137c" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.390382 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" event={"ID":"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e","Type":"ContainerStarted","Data":"cd857fd81b9d60885f23abd3b2c07fd82f462c02c36ab738b3626ed94de1c59d"} Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.391417 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.430499 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" podStartSLOduration=3.430481457 podStartE2EDuration="3.430481457s" podCreationTimestamp="2026-02-20 08:22:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:22:13.42453272 +0000 UTC m=+982.399027550" watchObservedRunningTime="2026-02-20 08:22:13.430481457 +0000 UTC m=+982.404976277" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.686129 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-v6srk"] Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.699164 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-g72bc"] Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.700422 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.704804 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.706884 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.707096 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.707287 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.707396 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-v9rj2" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.734509 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-g72bc"] Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.765941 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-fernet-keys\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.766028 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-combined-ca-bundle\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.766104 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-scripts\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.766130 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-credential-keys\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.766212 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-config-data\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.766235 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p27c6\" (UniqueName: \"kubernetes.io/projected/9959c3d4-fb0e-473c-8bbe-065110a3e03a-kube-api-access-p27c6\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.797202 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lnp7g" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.800211 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lnp7g" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.855084 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-gmdvb"] Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.862325 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.871848 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-scripts\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.871892 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-credential-keys\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.871987 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-config-data\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.872002 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p27c6\" (UniqueName: \"kubernetes.io/projected/9959c3d4-fb0e-473c-8bbe-065110a3e03a-kube-api-access-p27c6\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.872059 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-fernet-keys\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.872083 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-combined-ca-bundle\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.895572 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-credential-keys\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.904628 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-config-data\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.909268 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-fernet-keys\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.912447 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-combined-ca-bundle\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.925028 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-gmdvb"] Feb 20 08:22:13 crc kubenswrapper[4948]: W0220 08:22:13.925927 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7960b704_21ac_4d75_8226_a9bbae4c2b50.slice/crio-e3cb32ae5e929563597aaafd43d3401d85066f36163aa0afc9720c6e975f05be WatchSource:0}: Error finding container e3cb32ae5e929563597aaafd43d3401d85066f36163aa0afc9720c6e975f05be: Status 404 returned error can't find the container with id e3cb32ae5e929563597aaafd43d3401d85066f36163aa0afc9720c6e975f05be Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.942320 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-scripts\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.944902 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-v6srk"] Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.972352 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p27c6\" (UniqueName: \"kubernetes.io/projected/9959c3d4-fb0e-473c-8bbe-065110a3e03a-kube-api-access-p27c6\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.973960 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-config\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.974083 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.974145 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.974165 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-dns-svc\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.974213 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28ln7\" (UniqueName: \"kubernetes.io/projected/69c38b92-a051-4834-bf34-8612edb31dfa-kube-api-access-28ln7\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.974229 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.979257 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lnp7g" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.018862 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-g5zxh"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.019883 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-g5zxh" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.057006 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7d578c747f-8f6hc"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.058348 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.066760 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.066944 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.067095 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-4txrz" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.067270 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.067482 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.067620 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-75xnq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.067846 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.068266 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.081833 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.081887 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-dns-svc\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.081956 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb362d0e-68b4-4d48-885c-a6894ac71e6e-combined-ca-bundle\") pod \"neutron-db-sync-g5zxh\" (UID: \"bb362d0e-68b4-4d48-885c-a6894ac71e6e\") " pod="openstack/neutron-db-sync-g5zxh" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.082038 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bb362d0e-68b4-4d48-885c-a6894ac71e6e-config\") pod \"neutron-db-sync-g5zxh\" (UID: \"bb362d0e-68b4-4d48-885c-a6894ac71e6e\") " pod="openstack/neutron-db-sync-g5zxh" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.082265 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28ln7\" (UniqueName: \"kubernetes.io/projected/69c38b92-a051-4834-bf34-8612edb31dfa-kube-api-access-28ln7\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.082291 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.082379 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npm75\" (UniqueName: \"kubernetes.io/projected/bb362d0e-68b4-4d48-885c-a6894ac71e6e-kube-api-access-npm75\") pod \"neutron-db-sync-g5zxh\" (UID: \"bb362d0e-68b4-4d48-885c-a6894ac71e6e\") " pod="openstack/neutron-db-sync-g5zxh" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.082437 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-config\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.082456 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.083225 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-dns-svc\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.083795 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.094766 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.094769 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.095469 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-config\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.099423 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-g5zxh"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.166074 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d578c747f-8f6hc"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.192344 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-scripts\") pod \"horizon-7d578c747f-8f6hc\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.192675 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npm75\" (UniqueName: \"kubernetes.io/projected/bb362d0e-68b4-4d48-885c-a6894ac71e6e-kube-api-access-npm75\") pod \"neutron-db-sync-g5zxh\" (UID: \"bb362d0e-68b4-4d48-885c-a6894ac71e6e\") " pod="openstack/neutron-db-sync-g5zxh" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.192695 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-logs\") pod \"horizon-7d578c747f-8f6hc\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.192735 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrsxt\" (UniqueName: \"kubernetes.io/projected/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-kube-api-access-jrsxt\") pod \"horizon-7d578c747f-8f6hc\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.192780 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-horizon-secret-key\") pod \"horizon-7d578c747f-8f6hc\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.192857 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb362d0e-68b4-4d48-885c-a6894ac71e6e-combined-ca-bundle\") pod \"neutron-db-sync-g5zxh\" (UID: \"bb362d0e-68b4-4d48-885c-a6894ac71e6e\") " pod="openstack/neutron-db-sync-g5zxh" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.192872 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bb362d0e-68b4-4d48-885c-a6894ac71e6e-config\") pod \"neutron-db-sync-g5zxh\" (UID: \"bb362d0e-68b4-4d48-885c-a6894ac71e6e\") " pod="openstack/neutron-db-sync-g5zxh" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.192938 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-config-data\") pod \"horizon-7d578c747f-8f6hc\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.202244 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb362d0e-68b4-4d48-885c-a6894ac71e6e-combined-ca-bundle\") pod \"neutron-db-sync-g5zxh\" (UID: \"bb362d0e-68b4-4d48-885c-a6894ac71e6e\") " pod="openstack/neutron-db-sync-g5zxh" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.207627 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28ln7\" (UniqueName: \"kubernetes.io/projected/69c38b92-a051-4834-bf34-8612edb31dfa-kube-api-access-28ln7\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.210083 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-xt2zq"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.226875 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/bb362d0e-68b4-4d48-885c-a6894ac71e6e-config\") pod \"neutron-db-sync-g5zxh\" (UID: \"bb362d0e-68b4-4d48-885c-a6894ac71e6e\") " pod="openstack/neutron-db-sync-g5zxh" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.231166 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.241292 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-q6pbd" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.241475 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.246215 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.255949 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-xt2zq"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.265517 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npm75\" (UniqueName: \"kubernetes.io/projected/bb362d0e-68b4-4d48-885c-a6894ac71e6e-kube-api-access-npm75\") pod \"neutron-db-sync-g5zxh\" (UID: \"bb362d0e-68b4-4d48-885c-a6894ac71e6e\") " pod="openstack/neutron-db-sync-g5zxh" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.268036 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.270074 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.280477 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.280702 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.294894 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f706b741-86f9-49a3-95d5-85ee62eb3668-etc-machine-id\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.294937 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-config-data\") pod \"horizon-7d578c747f-8f6hc\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.294997 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-scripts\") pod \"horizon-7d578c747f-8f6hc\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.295017 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-logs\") pod \"horizon-7d578c747f-8f6hc\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.295044 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrsxt\" (UniqueName: \"kubernetes.io/projected/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-kube-api-access-jrsxt\") pod \"horizon-7d578c747f-8f6hc\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.295075 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-horizon-secret-key\") pod \"horizon-7d578c747f-8f6hc\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.295109 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-config-data\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.295132 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-db-sync-config-data\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.295153 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-combined-ca-bundle\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.295168 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-scripts\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.295231 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwtv5\" (UniqueName: \"kubernetes.io/projected/f706b741-86f9-49a3-95d5-85ee62eb3668-kube-api-access-fwtv5\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.296342 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-config-data\") pod \"horizon-7d578c747f-8f6hc\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.300862 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.302036 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-logs\") pod \"horizon-7d578c747f-8f6hc\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.310350 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-scripts\") pod \"horizon-7d578c747f-8f6hc\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.330463 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-horizon-secret-key\") pod \"horizon-7d578c747f-8f6hc\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.363921 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrsxt\" (UniqueName: \"kubernetes.io/projected/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-kube-api-access-jrsxt\") pod \"horizon-7d578c747f-8f6hc\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.387032 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.388300 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.402011 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-config-data\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.402090 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.402134 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-config-data\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.402166 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-db-sync-config-data\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.402187 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-combined-ca-bundle\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.402208 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-scripts\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.402262 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwtv5\" (UniqueName: \"kubernetes.io/projected/f706b741-86f9-49a3-95d5-85ee62eb3668-kube-api-access-fwtv5\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.402299 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f706b741-86f9-49a3-95d5-85ee62eb3668-etc-machine-id\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.402331 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.402365 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-scripts\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.402383 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77njn\" (UniqueName: \"kubernetes.io/projected/89b49c50-bae7-47fe-b85f-7a718953933d-kube-api-access-77njn\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.402399 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89b49c50-bae7-47fe-b85f-7a718953933d-run-httpd\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.402423 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89b49c50-bae7-47fe-b85f-7a718953933d-log-httpd\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.403329 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-6mz7x"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.404761 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.405791 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.408093 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f706b741-86f9-49a3-95d5-85ee62eb3668-etc-machine-id\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.408601 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-bgxj8" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.409149 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.417260 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.423073 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-gmdvb"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.432772 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-2fq52" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.432995 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.433105 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.433566 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-db-sync-config-data\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.434117 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-config-data\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.437605 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-combined-ca-bundle\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.440947 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-scripts\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.473391 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-g5zxh" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.478400 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwtv5\" (UniqueName: \"kubernetes.io/projected/f706b741-86f9-49a3-95d5-85ee62eb3668-kube-api-access-fwtv5\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.478476 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.488778 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.525120 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-6mz7x"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.529552 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.529674 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1227b752-8a3c-4c01-91bf-0662c1b1a231-logs\") pod \"placement-db-sync-6mz7x\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.529763 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.529802 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b8ad7a10-6353-45f2-a11d-126d0d4786b0-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.529872 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-config-data\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.529984 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8ad7a10-6353-45f2-a11d-126d0d4786b0-logs\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.530018 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.530062 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rtp8\" (UniqueName: \"kubernetes.io/projected/1227b752-8a3c-4c01-91bf-0662c1b1a231-kube-api-access-5rtp8\") pod \"placement-db-sync-6mz7x\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.530094 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqhps\" (UniqueName: \"kubernetes.io/projected/b8ad7a10-6353-45f2-a11d-126d0d4786b0-kube-api-access-kqhps\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.530126 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-scripts\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.530149 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77njn\" (UniqueName: \"kubernetes.io/projected/89b49c50-bae7-47fe-b85f-7a718953933d-kube-api-access-77njn\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.530173 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89b49c50-bae7-47fe-b85f-7a718953933d-run-httpd\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.530214 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-scripts\") pod \"placement-db-sync-6mz7x\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.530238 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.530268 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89b49c50-bae7-47fe-b85f-7a718953933d-log-httpd\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.530316 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-config-data\") pod \"placement-db-sync-6mz7x\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.530358 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-combined-ca-bundle\") pod \"placement-db-sync-6mz7x\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.530391 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-config-data\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.530415 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-scripts\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.531703 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89b49c50-bae7-47fe-b85f-7a718953933d-run-httpd\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.542489 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-scripts\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.544520 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" event={"ID":"7960b704-21ac-4d75-8226-a9bbae4c2b50","Type":"ContainerStarted","Data":"e3cb32ae5e929563597aaafd43d3401d85066f36163aa0afc9720c6e975f05be"} Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.544823 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" podUID="0b8788ed-d6f3-4344-b8b2-7bb554ecd25e" containerName="dnsmasq-dns" containerID="cri-o://cd857fd81b9d60885f23abd3b2c07fd82f462c02c36ab738b3626ed94de1c59d" gracePeriod=10 Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.548157 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89b49c50-bae7-47fe-b85f-7a718953933d-log-httpd\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.551150 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-config-data\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.554903 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77njn\" (UniqueName: \"kubernetes.io/projected/89b49c50-bae7-47fe-b85f-7a718953933d-kube-api-access-77njn\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.560201 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.563509 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.564114 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-c7zzv"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.565813 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-c7zzv" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.567984 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-8rh2b" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.568152 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.573235 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-54d5f75fb9-s2659"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.574543 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.583966 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.588060 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-c7zzv"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.602375 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-96668"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.604308 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.605027 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.614329 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-54d5f75fb9-s2659"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.621264 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-96668"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636446 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1227b752-8a3c-4c01-91bf-0662c1b1a231-logs\") pod \"placement-db-sync-6mz7x\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636497 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/549beb55-bab4-4b1a-bf98-48de2e12956f-config-data\") pod \"horizon-54d5f75fb9-s2659\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636516 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636535 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/549beb55-bab4-4b1a-bf98-48de2e12956f-horizon-secret-key\") pod \"horizon-54d5f75fb9-s2659\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636553 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b8ad7a10-6353-45f2-a11d-126d0d4786b0-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636575 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/549beb55-bab4-4b1a-bf98-48de2e12956f-scripts\") pod \"horizon-54d5f75fb9-s2659\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636594 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-config-data\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636620 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/549beb55-bab4-4b1a-bf98-48de2e12956f-logs\") pod \"horizon-54d5f75fb9-s2659\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636637 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f1f108e1-27bb-433e-b7e1-f9cc7f778182-db-sync-config-data\") pod \"barbican-db-sync-c7zzv\" (UID: \"f1f108e1-27bb-433e-b7e1-f9cc7f778182\") " pod="openstack/barbican-db-sync-c7zzv" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636657 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8ad7a10-6353-45f2-a11d-126d0d4786b0-logs\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636671 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxqmm\" (UniqueName: \"kubernetes.io/projected/549beb55-bab4-4b1a-bf98-48de2e12956f-kube-api-access-gxqmm\") pod \"horizon-54d5f75fb9-s2659\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636694 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmlgz\" (UniqueName: \"kubernetes.io/projected/f1f108e1-27bb-433e-b7e1-f9cc7f778182-kube-api-access-gmlgz\") pod \"barbican-db-sync-c7zzv\" (UID: \"f1f108e1-27bb-433e-b7e1-f9cc7f778182\") " pod="openstack/barbican-db-sync-c7zzv" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636713 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rtp8\" (UniqueName: \"kubernetes.io/projected/1227b752-8a3c-4c01-91bf-0662c1b1a231-kube-api-access-5rtp8\") pod \"placement-db-sync-6mz7x\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636730 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqhps\" (UniqueName: \"kubernetes.io/projected/b8ad7a10-6353-45f2-a11d-126d0d4786b0-kube-api-access-kqhps\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636753 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1f108e1-27bb-433e-b7e1-f9cc7f778182-combined-ca-bundle\") pod \"barbican-db-sync-c7zzv\" (UID: \"f1f108e1-27bb-433e-b7e1-f9cc7f778182\") " pod="openstack/barbican-db-sync-c7zzv" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636770 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-scripts\") pod \"placement-db-sync-6mz7x\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636787 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636810 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-config-data\") pod \"placement-db-sync-6mz7x\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636830 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-combined-ca-bundle\") pod \"placement-db-sync-6mz7x\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636856 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-scripts\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.641536 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1227b752-8a3c-4c01-91bf-0662c1b1a231-logs\") pod \"placement-db-sync-6mz7x\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.642256 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.642574 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8ad7a10-6353-45f2-a11d-126d0d4786b0-logs\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.649497 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b8ad7a10-6353-45f2-a11d-126d0d4786b0-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.652451 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-config-data\") pod \"placement-db-sync-6mz7x\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.655158 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.664556 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-scripts\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.668655 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-combined-ca-bundle\") pod \"placement-db-sync-6mz7x\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.672600 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqhps\" (UniqueName: \"kubernetes.io/projected/b8ad7a10-6353-45f2-a11d-126d0d4786b0-kube-api-access-kqhps\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.677352 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-scripts\") pod \"placement-db-sync-6mz7x\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.678330 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-config-data\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.679344 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rtp8\" (UniqueName: \"kubernetes.io/projected/1227b752-8a3c-4c01-91bf-0662c1b1a231-kube-api-access-5rtp8\") pod \"placement-db-sync-6mz7x\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.688587 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.690927 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.695754 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.705923 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.731402 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.742671 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-config\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.742746 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.742783 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.742824 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.742847 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.743147 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/549beb55-bab4-4b1a-bf98-48de2e12956f-config-data\") pod \"horizon-54d5f75fb9-s2659\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.743171 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/549beb55-bab4-4b1a-bf98-48de2e12956f-horizon-secret-key\") pod \"horizon-54d5f75fb9-s2659\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.743216 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.743238 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/549beb55-bab4-4b1a-bf98-48de2e12956f-scripts\") pod \"horizon-54d5f75fb9-s2659\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.743291 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.743310 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/549beb55-bab4-4b1a-bf98-48de2e12956f-logs\") pod \"horizon-54d5f75fb9-s2659\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.743327 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f1f108e1-27bb-433e-b7e1-f9cc7f778182-db-sync-config-data\") pod \"barbican-db-sync-c7zzv\" (UID: \"f1f108e1-27bb-433e-b7e1-f9cc7f778182\") " pod="openstack/barbican-db-sync-c7zzv" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.743343 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6e945937-1e9d-465b-a68f-c755119ab6f9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.743364 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxqmm\" (UniqueName: \"kubernetes.io/projected/549beb55-bab4-4b1a-bf98-48de2e12956f-kube-api-access-gxqmm\") pod \"horizon-54d5f75fb9-s2659\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.743387 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svj9w\" (UniqueName: \"kubernetes.io/projected/6e945937-1e9d-465b-a68f-c755119ab6f9-kube-api-access-svj9w\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.743406 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmlgz\" (UniqueName: \"kubernetes.io/projected/f1f108e1-27bb-433e-b7e1-f9cc7f778182-kube-api-access-gmlgz\") pod \"barbican-db-sync-c7zzv\" (UID: \"f1f108e1-27bb-433e-b7e1-f9cc7f778182\") " pod="openstack/barbican-db-sync-c7zzv" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.743427 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqddr\" (UniqueName: \"kubernetes.io/projected/94b82d55-b467-408f-8eb1-6da4648dde97-kube-api-access-rqddr\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.743445 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1f108e1-27bb-433e-b7e1-f9cc7f778182-combined-ca-bundle\") pod \"barbican-db-sync-c7zzv\" (UID: \"f1f108e1-27bb-433e-b7e1-f9cc7f778182\") " pod="openstack/barbican-db-sync-c7zzv" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.743460 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.743477 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.743497 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e945937-1e9d-465b-a68f-c755119ab6f9-logs\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.745796 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/549beb55-bab4-4b1a-bf98-48de2e12956f-config-data\") pod \"horizon-54d5f75fb9-s2659\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.748302 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/549beb55-bab4-4b1a-bf98-48de2e12956f-scripts\") pod \"horizon-54d5f75fb9-s2659\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.748905 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/549beb55-bab4-4b1a-bf98-48de2e12956f-logs\") pod \"horizon-54d5f75fb9-s2659\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.775488 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.777454 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/549beb55-bab4-4b1a-bf98-48de2e12956f-horizon-secret-key\") pod \"horizon-54d5f75fb9-s2659\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.778067 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1f108e1-27bb-433e-b7e1-f9cc7f778182-combined-ca-bundle\") pod \"barbican-db-sync-c7zzv\" (UID: \"f1f108e1-27bb-433e-b7e1-f9cc7f778182\") " pod="openstack/barbican-db-sync-c7zzv" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.787172 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmlgz\" (UniqueName: \"kubernetes.io/projected/f1f108e1-27bb-433e-b7e1-f9cc7f778182-kube-api-access-gmlgz\") pod \"barbican-db-sync-c7zzv\" (UID: \"f1f108e1-27bb-433e-b7e1-f9cc7f778182\") " pod="openstack/barbican-db-sync-c7zzv" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.792683 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.793882 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f1f108e1-27bb-433e-b7e1-f9cc7f778182-db-sync-config-data\") pod \"barbican-db-sync-c7zzv\" (UID: \"f1f108e1-27bb-433e-b7e1-f9cc7f778182\") " pod="openstack/barbican-db-sync-c7zzv" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.796466 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxqmm\" (UniqueName: \"kubernetes.io/projected/549beb55-bab4-4b1a-bf98-48de2e12956f-kube-api-access-gxqmm\") pod \"horizon-54d5f75fb9-s2659\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.844867 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.844959 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.845007 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6e945937-1e9d-465b-a68f-c755119ab6f9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.845036 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svj9w\" (UniqueName: \"kubernetes.io/projected/6e945937-1e9d-465b-a68f-c755119ab6f9-kube-api-access-svj9w\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.845103 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqddr\" (UniqueName: \"kubernetes.io/projected/94b82d55-b467-408f-8eb1-6da4648dde97-kube-api-access-rqddr\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.845122 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.845168 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.845193 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e945937-1e9d-465b-a68f-c755119ab6f9-logs\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.845261 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.845282 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-config\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.845342 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.845366 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.845407 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.846793 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.851915 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.852212 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6e945937-1e9d-465b-a68f-c755119ab6f9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.852636 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e945937-1e9d-465b-a68f-c755119ab6f9-logs\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.853229 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.853475 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.853747 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.856771 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-config\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.860427 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.865929 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.874919 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqddr\" (UniqueName: \"kubernetes.io/projected/94b82d55-b467-408f-8eb1-6da4648dde97-kube-api-access-rqddr\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.875572 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.884948 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svj9w\" (UniqueName: \"kubernetes.io/projected/6e945937-1e9d-465b-a68f-c755119ab6f9-kube-api-access-svj9w\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.924846 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.931497 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-c7zzv" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.954043 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.968221 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.989095 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-g72bc"] Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.032774 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.156569 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-gmdvb"] Feb 20 08:22:15 crc kubenswrapper[4948]: W0220 08:22:15.214770 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69c38b92_a051_4834_bf34_8612edb31dfa.slice/crio-4750fdb6550fc9d53336366b666ece8cbabd6c85ec2d4df49fece59f3bc13a6a WatchSource:0}: Error finding container 4750fdb6550fc9d53336366b666ece8cbabd6c85ec2d4df49fece59f3bc13a6a: Status 404 returned error can't find the container with id 4750fdb6550fc9d53336366b666ece8cbabd6c85ec2d4df49fece59f3bc13a6a Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.218776 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d578c747f-8f6hc"] Feb 20 08:22:15 crc kubenswrapper[4948]: W0220 08:22:15.225347 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8425fd12_ae45_4e7e_af4d_9bfc3c3d0eb4.slice/crio-3f991d1b028cfdcb1512441c7603f82fee2c2bba86ea276855760a70987a1d8f WatchSource:0}: Error finding container 3f991d1b028cfdcb1512441c7603f82fee2c2bba86ea276855760a70987a1d8f: Status 404 returned error can't find the container with id 3f991d1b028cfdcb1512441c7603f82fee2c2bba86ea276855760a70987a1d8f Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.453909 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-g5zxh"] Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.464136 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-xt2zq"] Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.576024 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d578c747f-8f6hc" event={"ID":"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4","Type":"ContainerStarted","Data":"3f991d1b028cfdcb1512441c7603f82fee2c2bba86ea276855760a70987a1d8f"} Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.585775 4948 generic.go:334] "Generic (PLEG): container finished" podID="7960b704-21ac-4d75-8226-a9bbae4c2b50" containerID="f7f0bf783e5f62b8ff1fccd63b4e33b097f4a17ad1c512a814c4baa9c6f1eb03" exitCode=0 Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.586038 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" event={"ID":"7960b704-21ac-4d75-8226-a9bbae4c2b50","Type":"ContainerDied","Data":"f7f0bf783e5f62b8ff1fccd63b4e33b097f4a17ad1c512a814c4baa9c6f1eb03"} Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.588090 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.593460 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-g72bc" event={"ID":"9959c3d4-fb0e-473c-8bbe-065110a3e03a","Type":"ContainerStarted","Data":"fa275de1b67d8651c437fb815c4f8470dc8d77cd9885aaddee1a7419f33fd25c"} Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.593497 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-g72bc" event={"ID":"9959c3d4-fb0e-473c-8bbe-065110a3e03a","Type":"ContainerStarted","Data":"1eedce1f0a19f6576c482c3583bf052a9786f8dcd27e5bce2d34fd0ef16f70ad"} Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.610809 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-xt2zq" event={"ID":"f706b741-86f9-49a3-95d5-85ee62eb3668","Type":"ContainerStarted","Data":"571bfb34e93031079bb93437b09bfef171c7529235d603237690b075a9933988"} Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.615581 4948 generic.go:334] "Generic (PLEG): container finished" podID="0b8788ed-d6f3-4344-b8b2-7bb554ecd25e" containerID="cd857fd81b9d60885f23abd3b2c07fd82f462c02c36ab738b3626ed94de1c59d" exitCode=0 Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.615632 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" event={"ID":"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e","Type":"ContainerDied","Data":"cd857fd81b9d60885f23abd3b2c07fd82f462c02c36ab738b3626ed94de1c59d"} Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.615656 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" event={"ID":"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e","Type":"ContainerDied","Data":"5b5cca04b9c3f166250baa067797237d67a3b5df1ae92815a7da886a955a6658"} Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.615674 4948 scope.go:117] "RemoveContainer" containerID="cd857fd81b9d60885f23abd3b2c07fd82f462c02c36ab738b3626ed94de1c59d" Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.615781 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.643871 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-g5zxh" event={"ID":"bb362d0e-68b4-4d48-885c-a6894ac71e6e","Type":"ContainerStarted","Data":"fb0ac2219219fe66e5a42c9ab8b34f9c1f674fc235bce862e34dedbfa83dd480"} Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.665706 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" event={"ID":"69c38b92-a051-4834-bf34-8612edb31dfa","Type":"ContainerStarted","Data":"545ede59ed041c4b028c3e626eff3b4e85ad826a2ef8b2830cd52655853e1ad4"} Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.665744 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" event={"ID":"69c38b92-a051-4834-bf34-8612edb31dfa","Type":"ContainerStarted","Data":"4750fdb6550fc9d53336366b666ece8cbabd6c85ec2d4df49fece59f3bc13a6a"} Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.678052 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-dns-svc\") pod \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.678109 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-config\") pod \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.678177 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57xf5\" (UniqueName: \"kubernetes.io/projected/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-kube-api-access-57xf5\") pod \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.678237 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-dns-swift-storage-0\") pod \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.678315 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-ovsdbserver-sb\") pod \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.678344 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-ovsdbserver-nb\") pod \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.684386 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.691885 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-kube-api-access-57xf5" (OuterVolumeSpecName: "kube-api-access-57xf5") pod "0b8788ed-d6f3-4344-b8b2-7bb554ecd25e" (UID: "0b8788ed-d6f3-4344-b8b2-7bb554ecd25e"). InnerVolumeSpecName "kube-api-access-57xf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.725004 4948 scope.go:117] "RemoveContainer" containerID="596bfc5f705d539b55d6de0f0ebc270b6e2a7ef5c99c535a13f5bc566d754f58" Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.744658 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-g72bc" podStartSLOduration=2.744634106 podStartE2EDuration="2.744634106s" podCreationTimestamp="2026-02-20 08:22:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:22:15.651666045 +0000 UTC m=+984.626160865" watchObservedRunningTime="2026-02-20 08:22:15.744634106 +0000 UTC m=+984.719128926" Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.783130 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57xf5\" (UniqueName: \"kubernetes.io/projected/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-kube-api-access-57xf5\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.789361 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.813381 4948 scope.go:117] "RemoveContainer" containerID="cd857fd81b9d60885f23abd3b2c07fd82f462c02c36ab738b3626ed94de1c59d" Feb 20 08:22:15 crc kubenswrapper[4948]: E0220 08:22:15.816915 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd857fd81b9d60885f23abd3b2c07fd82f462c02c36ab738b3626ed94de1c59d\": container with ID starting with cd857fd81b9d60885f23abd3b2c07fd82f462c02c36ab738b3626ed94de1c59d not found: ID does not exist" containerID="cd857fd81b9d60885f23abd3b2c07fd82f462c02c36ab738b3626ed94de1c59d" Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.816949 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd857fd81b9d60885f23abd3b2c07fd82f462c02c36ab738b3626ed94de1c59d"} err="failed to get container status \"cd857fd81b9d60885f23abd3b2c07fd82f462c02c36ab738b3626ed94de1c59d\": rpc error: code = NotFound desc = could not find container \"cd857fd81b9d60885f23abd3b2c07fd82f462c02c36ab738b3626ed94de1c59d\": container with ID starting with cd857fd81b9d60885f23abd3b2c07fd82f462c02c36ab738b3626ed94de1c59d not found: ID does not exist" Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.816986 4948 scope.go:117] "RemoveContainer" containerID="596bfc5f705d539b55d6de0f0ebc270b6e2a7ef5c99c535a13f5bc566d754f58" Feb 20 08:22:15 crc kubenswrapper[4948]: E0220 08:22:15.821125 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"596bfc5f705d539b55d6de0f0ebc270b6e2a7ef5c99c535a13f5bc566d754f58\": container with ID starting with 596bfc5f705d539b55d6de0f0ebc270b6e2a7ef5c99c535a13f5bc566d754f58 not found: ID does not exist" containerID="596bfc5f705d539b55d6de0f0ebc270b6e2a7ef5c99c535a13f5bc566d754f58" Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.821175 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"596bfc5f705d539b55d6de0f0ebc270b6e2a7ef5c99c535a13f5bc566d754f58"} err="failed to get container status \"596bfc5f705d539b55d6de0f0ebc270b6e2a7ef5c99c535a13f5bc566d754f58\": rpc error: code = NotFound desc = could not find container \"596bfc5f705d539b55d6de0f0ebc270b6e2a7ef5c99c535a13f5bc566d754f58\": container with ID starting with 596bfc5f705d539b55d6de0f0ebc270b6e2a7ef5c99c535a13f5bc566d754f58 not found: ID does not exist" Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.840432 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lnp7g" Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.912823 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lnp7g"] Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.964519 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0b8788ed-d6f3-4344-b8b2-7bb554ecd25e" (UID: "0b8788ed-d6f3-4344-b8b2-7bb554ecd25e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.002760 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-config" (OuterVolumeSpecName: "config") pod "0b8788ed-d6f3-4344-b8b2-7bb554ecd25e" (UID: "0b8788ed-d6f3-4344-b8b2-7bb554ecd25e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.002869 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.019816 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0b8788ed-d6f3-4344-b8b2-7bb554ecd25e" (UID: "0b8788ed-d6f3-4344-b8b2-7bb554ecd25e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.031454 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0b8788ed-d6f3-4344-b8b2-7bb554ecd25e" (UID: "0b8788ed-d6f3-4344-b8b2-7bb554ecd25e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.034660 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0b8788ed-d6f3-4344-b8b2-7bb554ecd25e" (UID: "0b8788ed-d6f3-4344-b8b2-7bb554ecd25e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.110531 4948 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.110567 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.110576 4948 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.110585 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.174024 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.218501 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7d578c747f-8f6hc"] Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.251592 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-68c9c4f5fc-2vtb4"] Feb 20 08:22:16 crc kubenswrapper[4948]: E0220 08:22:16.251964 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b8788ed-d6f3-4344-b8b2-7bb554ecd25e" containerName="dnsmasq-dns" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.251989 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b8788ed-d6f3-4344-b8b2-7bb554ecd25e" containerName="dnsmasq-dns" Feb 20 08:22:16 crc kubenswrapper[4948]: E0220 08:22:16.252009 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b8788ed-d6f3-4344-b8b2-7bb554ecd25e" containerName="init" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.252016 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b8788ed-d6f3-4344-b8b2-7bb554ecd25e" containerName="init" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.253167 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b8788ed-d6f3-4344-b8b2-7bb554ecd25e" containerName="dnsmasq-dns" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.254091 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.278925 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-68c9c4f5fc-2vtb4"] Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.286404 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-mvrcc"] Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.295036 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-mvrcc"] Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.315451 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.321912 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/136dd621-a5c5-481a-9b86-66f8cb72486d-config-data\") pod \"horizon-68c9c4f5fc-2vtb4\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.321988 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/136dd621-a5c5-481a-9b86-66f8cb72486d-logs\") pod \"horizon-68c9c4f5fc-2vtb4\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.322015 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/136dd621-a5c5-481a-9b86-66f8cb72486d-horizon-secret-key\") pod \"horizon-68c9c4f5fc-2vtb4\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.322072 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vp59z\" (UniqueName: \"kubernetes.io/projected/136dd621-a5c5-481a-9b86-66f8cb72486d-kube-api-access-vp59z\") pod \"horizon-68c9c4f5fc-2vtb4\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.322090 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/136dd621-a5c5-481a-9b86-66f8cb72486d-scripts\") pod \"horizon-68c9c4f5fc-2vtb4\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.428929 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/136dd621-a5c5-481a-9b86-66f8cb72486d-config-data\") pod \"horizon-68c9c4f5fc-2vtb4\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.429029 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/136dd621-a5c5-481a-9b86-66f8cb72486d-logs\") pod \"horizon-68c9c4f5fc-2vtb4\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.429067 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/136dd621-a5c5-481a-9b86-66f8cb72486d-horizon-secret-key\") pod \"horizon-68c9c4f5fc-2vtb4\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.429136 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vp59z\" (UniqueName: \"kubernetes.io/projected/136dd621-a5c5-481a-9b86-66f8cb72486d-kube-api-access-vp59z\") pod \"horizon-68c9c4f5fc-2vtb4\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.429159 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/136dd621-a5c5-481a-9b86-66f8cb72486d-scripts\") pod \"horizon-68c9c4f5fc-2vtb4\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.429504 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/136dd621-a5c5-481a-9b86-66f8cb72486d-logs\") pod \"horizon-68c9c4f5fc-2vtb4\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.430106 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/136dd621-a5c5-481a-9b86-66f8cb72486d-scripts\") pod \"horizon-68c9c4f5fc-2vtb4\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.430376 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/136dd621-a5c5-481a-9b86-66f8cb72486d-config-data\") pod \"horizon-68c9c4f5fc-2vtb4\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.466809 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/136dd621-a5c5-481a-9b86-66f8cb72486d-horizon-secret-key\") pod \"horizon-68c9c4f5fc-2vtb4\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.473025 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vp59z\" (UniqueName: \"kubernetes.io/projected/136dd621-a5c5-481a-9b86-66f8cb72486d-kube-api-access-vp59z\") pod \"horizon-68c9c4f5fc-2vtb4\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.514679 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.527863 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.556885 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.570928 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-c7zzv"] Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.591871 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-54d5f75fb9-s2659"] Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.604211 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.621530 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-96668"] Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.636741 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-dns-svc\") pod \"7960b704-21ac-4d75-8226-a9bbae4c2b50\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.636921 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-dns-svc\") pod \"69c38b92-a051-4834-bf34-8612edb31dfa\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.636958 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28ln7\" (UniqueName: \"kubernetes.io/projected/69c38b92-a051-4834-bf34-8612edb31dfa-kube-api-access-28ln7\") pod \"69c38b92-a051-4834-bf34-8612edb31dfa\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.639315 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-ovsdbserver-sb\") pod \"7960b704-21ac-4d75-8226-a9bbae4c2b50\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.639353 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-ovsdbserver-nb\") pod \"7960b704-21ac-4d75-8226-a9bbae4c2b50\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.639397 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-config\") pod \"7960b704-21ac-4d75-8226-a9bbae4c2b50\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.639440 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-ovsdbserver-nb\") pod \"69c38b92-a051-4834-bf34-8612edb31dfa\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.639469 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-dns-swift-storage-0\") pod \"69c38b92-a051-4834-bf34-8612edb31dfa\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.639491 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-ovsdbserver-sb\") pod \"69c38b92-a051-4834-bf34-8612edb31dfa\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.639534 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74rwq\" (UniqueName: \"kubernetes.io/projected/7960b704-21ac-4d75-8226-a9bbae4c2b50-kube-api-access-74rwq\") pod \"7960b704-21ac-4d75-8226-a9bbae4c2b50\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.639552 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-dns-swift-storage-0\") pod \"7960b704-21ac-4d75-8226-a9bbae4c2b50\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.639595 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-config\") pod \"69c38b92-a051-4834-bf34-8612edb31dfa\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " Feb 20 08:22:16 crc kubenswrapper[4948]: W0220 08:22:16.645724 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1227b752_8a3c_4c01_91bf_0662c1b1a231.slice/crio-5b0f5c257994604a44a9cf0e2c06abc610dcd1eb5c3d34cfd415bfd74671d08f WatchSource:0}: Error finding container 5b0f5c257994604a44a9cf0e2c06abc610dcd1eb5c3d34cfd415bfd74671d08f: Status 404 returned error can't find the container with id 5b0f5c257994604a44a9cf0e2c06abc610dcd1eb5c3d34cfd415bfd74671d08f Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.652593 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69c38b92-a051-4834-bf34-8612edb31dfa-kube-api-access-28ln7" (OuterVolumeSpecName: "kube-api-access-28ln7") pod "69c38b92-a051-4834-bf34-8612edb31dfa" (UID: "69c38b92-a051-4834-bf34-8612edb31dfa"). InnerVolumeSpecName "kube-api-access-28ln7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.673911 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-54d5f75fb9-s2659" event={"ID":"549beb55-bab4-4b1a-bf98-48de2e12956f","Type":"ContainerStarted","Data":"27986056f94b762ad4c0ec91cfd1562122b907c3b47800656945a8588571afe8"} Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.677013 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89b49c50-bae7-47fe-b85f-7a718953933d","Type":"ContainerStarted","Data":"2c29cb12a24e0769e2fc11186d9dff0648d1bcb261fcbab7060faff8fb76085a"} Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.684808 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-6mz7x"] Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.690851 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7960b704-21ac-4d75-8226-a9bbae4c2b50-kube-api-access-74rwq" (OuterVolumeSpecName: "kube-api-access-74rwq") pod "7960b704-21ac-4d75-8226-a9bbae4c2b50" (UID: "7960b704-21ac-4d75-8226-a9bbae4c2b50"). InnerVolumeSpecName "kube-api-access-74rwq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.702360 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-96668" event={"ID":"94b82d55-b467-408f-8eb1-6da4648dde97","Type":"ContainerStarted","Data":"af3aca3908294e0c1fe45deb6418a29e5d516dc4d54d54ff11acdb6845abd301"} Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.705002 4948 generic.go:334] "Generic (PLEG): container finished" podID="69c38b92-a051-4834-bf34-8612edb31dfa" containerID="545ede59ed041c4b028c3e626eff3b4e85ad826a2ef8b2830cd52655853e1ad4" exitCode=0 Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.705038 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" event={"ID":"69c38b92-a051-4834-bf34-8612edb31dfa","Type":"ContainerDied","Data":"545ede59ed041c4b028c3e626eff3b4e85ad826a2ef8b2830cd52655853e1ad4"} Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.705054 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" event={"ID":"69c38b92-a051-4834-bf34-8612edb31dfa","Type":"ContainerDied","Data":"4750fdb6550fc9d53336366b666ece8cbabd6c85ec2d4df49fece59f3bc13a6a"} Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.705069 4948 scope.go:117] "RemoveContainer" containerID="545ede59ed041c4b028c3e626eff3b4e85ad826a2ef8b2830cd52655853e1ad4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.705177 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.706839 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" event={"ID":"7960b704-21ac-4d75-8226-a9bbae4c2b50","Type":"ContainerDied","Data":"e3cb32ae5e929563597aaafd43d3401d85066f36163aa0afc9720c6e975f05be"} Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.706903 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.713245 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.714195 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b8ad7a10-6353-45f2-a11d-126d0d4786b0","Type":"ContainerStarted","Data":"434c4fc6de7f62028433ff88fb8c531a01a506655fd56b116472009d2d35dd9f"} Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.716405 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-g5zxh" event={"ID":"bb362d0e-68b4-4d48-885c-a6894ac71e6e","Type":"ContainerStarted","Data":"cb7241154041475439806eb631212532adb6e6775adf70e62bdf5cf4b6d19276"} Feb 20 08:22:16 crc kubenswrapper[4948]: W0220 08:22:16.724418 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e945937_1e9d_465b_a68f_c755119ab6f9.slice/crio-feaec65599576490d6ae435b0c4ffac29904876fc82b7f96fd44c9df4357c090 WatchSource:0}: Error finding container feaec65599576490d6ae435b0c4ffac29904876fc82b7f96fd44c9df4357c090: Status 404 returned error can't find the container with id feaec65599576490d6ae435b0c4ffac29904876fc82b7f96fd44c9df4357c090 Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.725574 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6mz7x" event={"ID":"1227b752-8a3c-4c01-91bf-0662c1b1a231","Type":"ContainerStarted","Data":"5b0f5c257994604a44a9cf0e2c06abc610dcd1eb5c3d34cfd415bfd74671d08f"} Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.731754 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-c7zzv" event={"ID":"f1f108e1-27bb-433e-b7e1-f9cc7f778182","Type":"ContainerStarted","Data":"1c7c890826add7e27d4d21f7cb357370f931db1e77090c60ac1c40f2ed6d57cc"} Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.735385 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-g5zxh" podStartSLOduration=3.735356072 podStartE2EDuration="3.735356072s" podCreationTimestamp="2026-02-20 08:22:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:22:16.731488736 +0000 UTC m=+985.705983556" watchObservedRunningTime="2026-02-20 08:22:16.735356072 +0000 UTC m=+985.709850892" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.745163 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28ln7\" (UniqueName: \"kubernetes.io/projected/69c38b92-a051-4834-bf34-8612edb31dfa-kube-api-access-28ln7\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.745195 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74rwq\" (UniqueName: \"kubernetes.io/projected/7960b704-21ac-4d75-8226-a9bbae4c2b50-kube-api-access-74rwq\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.754566 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-config" (OuterVolumeSpecName: "config") pod "69c38b92-a051-4834-bf34-8612edb31dfa" (UID: "69c38b92-a051-4834-bf34-8612edb31dfa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.764238 4948 scope.go:117] "RemoveContainer" containerID="545ede59ed041c4b028c3e626eff3b4e85ad826a2ef8b2830cd52655853e1ad4" Feb 20 08:22:16 crc kubenswrapper[4948]: E0220 08:22:16.766172 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"545ede59ed041c4b028c3e626eff3b4e85ad826a2ef8b2830cd52655853e1ad4\": container with ID starting with 545ede59ed041c4b028c3e626eff3b4e85ad826a2ef8b2830cd52655853e1ad4 not found: ID does not exist" containerID="545ede59ed041c4b028c3e626eff3b4e85ad826a2ef8b2830cd52655853e1ad4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.766203 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"545ede59ed041c4b028c3e626eff3b4e85ad826a2ef8b2830cd52655853e1ad4"} err="failed to get container status \"545ede59ed041c4b028c3e626eff3b4e85ad826a2ef8b2830cd52655853e1ad4\": rpc error: code = NotFound desc = could not find container \"545ede59ed041c4b028c3e626eff3b4e85ad826a2ef8b2830cd52655853e1ad4\": container with ID starting with 545ede59ed041c4b028c3e626eff3b4e85ad826a2ef8b2830cd52655853e1ad4 not found: ID does not exist" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.766221 4948 scope.go:117] "RemoveContainer" containerID="f7f0bf783e5f62b8ff1fccd63b4e33b097f4a17ad1c512a814c4baa9c6f1eb03" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.856997 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.894336 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "69c38b92-a051-4834-bf34-8612edb31dfa" (UID: "69c38b92-a051-4834-bf34-8612edb31dfa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.900352 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7960b704-21ac-4d75-8226-a9bbae4c2b50" (UID: "7960b704-21ac-4d75-8226-a9bbae4c2b50"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.909659 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7960b704-21ac-4d75-8226-a9bbae4c2b50" (UID: "7960b704-21ac-4d75-8226-a9bbae4c2b50"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.938043 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7960b704-21ac-4d75-8226-a9bbae4c2b50" (UID: "7960b704-21ac-4d75-8226-a9bbae4c2b50"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.941248 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "69c38b92-a051-4834-bf34-8612edb31dfa" (UID: "69c38b92-a051-4834-bf34-8612edb31dfa"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.943884 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "69c38b92-a051-4834-bf34-8612edb31dfa" (UID: "69c38b92-a051-4834-bf34-8612edb31dfa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.951930 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "69c38b92-a051-4834-bf34-8612edb31dfa" (UID: "69c38b92-a051-4834-bf34-8612edb31dfa"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.958326 4948 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.958346 4948 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.958354 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.958364 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.958372 4948 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.958380 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.958388 4948 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.960287 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-config" (OuterVolumeSpecName: "config") pod "7960b704-21ac-4d75-8226-a9bbae4c2b50" (UID: "7960b704-21ac-4d75-8226-a9bbae4c2b50"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.973483 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7960b704-21ac-4d75-8226-a9bbae4c2b50" (UID: "7960b704-21ac-4d75-8226-a9bbae4c2b50"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:17 crc kubenswrapper[4948]: I0220 08:22:17.059922 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:17 crc kubenswrapper[4948]: I0220 08:22:17.059954 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:17 crc kubenswrapper[4948]: I0220 08:22:17.195190 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-gmdvb"] Feb 20 08:22:17 crc kubenswrapper[4948]: I0220 08:22:17.243949 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-gmdvb"] Feb 20 08:22:17 crc kubenswrapper[4948]: I0220 08:22:17.266231 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-v6srk"] Feb 20 08:22:17 crc kubenswrapper[4948]: I0220 08:22:17.275082 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-v6srk"] Feb 20 08:22:17 crc kubenswrapper[4948]: I0220 08:22:17.281609 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-68c9c4f5fc-2vtb4"] Feb 20 08:22:17 crc kubenswrapper[4948]: E0220 08:22:17.513362 4948 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94b82d55_b467_408f_8eb1_6da4648dde97.slice/crio-bbaf3121e36eabf03661c450060f71b1a26f424ea932363040deddf76d19e0df.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94b82d55_b467_408f_8eb1_6da4648dde97.slice/crio-conmon-bbaf3121e36eabf03661c450060f71b1a26f424ea932363040deddf76d19e0df.scope\": RecentStats: unable to find data in memory cache]" Feb 20 08:22:17 crc kubenswrapper[4948]: I0220 08:22:17.743449 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b8788ed-d6f3-4344-b8b2-7bb554ecd25e" path="/var/lib/kubelet/pods/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e/volumes" Feb 20 08:22:17 crc kubenswrapper[4948]: I0220 08:22:17.744414 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69c38b92-a051-4834-bf34-8612edb31dfa" path="/var/lib/kubelet/pods/69c38b92-a051-4834-bf34-8612edb31dfa/volumes" Feb 20 08:22:17 crc kubenswrapper[4948]: I0220 08:22:17.745381 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7960b704-21ac-4d75-8226-a9bbae4c2b50" path="/var/lib/kubelet/pods/7960b704-21ac-4d75-8226-a9bbae4c2b50/volumes" Feb 20 08:22:17 crc kubenswrapper[4948]: I0220 08:22:17.789094 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b8ad7a10-6353-45f2-a11d-126d0d4786b0","Type":"ContainerStarted","Data":"fdfa9ea96d17b0fbac466add147012945b8f9c77bfe4c7e3939115948c20ff15"} Feb 20 08:22:17 crc kubenswrapper[4948]: I0220 08:22:17.797186 4948 generic.go:334] "Generic (PLEG): container finished" podID="94b82d55-b467-408f-8eb1-6da4648dde97" containerID="bbaf3121e36eabf03661c450060f71b1a26f424ea932363040deddf76d19e0df" exitCode=0 Feb 20 08:22:17 crc kubenswrapper[4948]: I0220 08:22:17.797270 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-96668" event={"ID":"94b82d55-b467-408f-8eb1-6da4648dde97","Type":"ContainerDied","Data":"bbaf3121e36eabf03661c450060f71b1a26f424ea932363040deddf76d19e0df"} Feb 20 08:22:17 crc kubenswrapper[4948]: I0220 08:22:17.804742 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6e945937-1e9d-465b-a68f-c755119ab6f9","Type":"ContainerStarted","Data":"feaec65599576490d6ae435b0c4ffac29904876fc82b7f96fd44c9df4357c090"} Feb 20 08:22:17 crc kubenswrapper[4948]: I0220 08:22:17.817451 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68c9c4f5fc-2vtb4" event={"ID":"136dd621-a5c5-481a-9b86-66f8cb72486d","Type":"ContainerStarted","Data":"45d6073bbc41d12c6fd4a83a0736dfc692eb915e2f643d37fcfcdd4574770ee8"} Feb 20 08:22:17 crc kubenswrapper[4948]: I0220 08:22:17.819008 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lnp7g" podUID="81e7d5de-2379-46c7-ba9b-ffcacad79e29" containerName="registry-server" containerID="cri-o://59eceb81df286be225188d13264d44596719c0a06ca1c39903c517b422013a68" gracePeriod=2 Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.495142 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lnp7g" Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.615470 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81e7d5de-2379-46c7-ba9b-ffcacad79e29-catalog-content\") pod \"81e7d5de-2379-46c7-ba9b-ffcacad79e29\" (UID: \"81e7d5de-2379-46c7-ba9b-ffcacad79e29\") " Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.615529 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81e7d5de-2379-46c7-ba9b-ffcacad79e29-utilities\") pod \"81e7d5de-2379-46c7-ba9b-ffcacad79e29\" (UID: \"81e7d5de-2379-46c7-ba9b-ffcacad79e29\") " Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.615551 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5vn9\" (UniqueName: \"kubernetes.io/projected/81e7d5de-2379-46c7-ba9b-ffcacad79e29-kube-api-access-w5vn9\") pod \"81e7d5de-2379-46c7-ba9b-ffcacad79e29\" (UID: \"81e7d5de-2379-46c7-ba9b-ffcacad79e29\") " Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.617143 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81e7d5de-2379-46c7-ba9b-ffcacad79e29-utilities" (OuterVolumeSpecName: "utilities") pod "81e7d5de-2379-46c7-ba9b-ffcacad79e29" (UID: "81e7d5de-2379-46c7-ba9b-ffcacad79e29"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.625716 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81e7d5de-2379-46c7-ba9b-ffcacad79e29-kube-api-access-w5vn9" (OuterVolumeSpecName: "kube-api-access-w5vn9") pod "81e7d5de-2379-46c7-ba9b-ffcacad79e29" (UID: "81e7d5de-2379-46c7-ba9b-ffcacad79e29"). InnerVolumeSpecName "kube-api-access-w5vn9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.650957 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81e7d5de-2379-46c7-ba9b-ffcacad79e29-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "81e7d5de-2379-46c7-ba9b-ffcacad79e29" (UID: "81e7d5de-2379-46c7-ba9b-ffcacad79e29"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.717973 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81e7d5de-2379-46c7-ba9b-ffcacad79e29-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.718238 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81e7d5de-2379-46c7-ba9b-ffcacad79e29-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.718255 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5vn9\" (UniqueName: \"kubernetes.io/projected/81e7d5de-2379-46c7-ba9b-ffcacad79e29-kube-api-access-w5vn9\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.837283 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lnp7g" Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.837584 4948 generic.go:334] "Generic (PLEG): container finished" podID="81e7d5de-2379-46c7-ba9b-ffcacad79e29" containerID="59eceb81df286be225188d13264d44596719c0a06ca1c39903c517b422013a68" exitCode=0 Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.837666 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnp7g" event={"ID":"81e7d5de-2379-46c7-ba9b-ffcacad79e29","Type":"ContainerDied","Data":"59eceb81df286be225188d13264d44596719c0a06ca1c39903c517b422013a68"} Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.837692 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnp7g" event={"ID":"81e7d5de-2379-46c7-ba9b-ffcacad79e29","Type":"ContainerDied","Data":"ac4ceb2acdc56fbdc4b52d619a840394196af858618bbcf55bc7c1dc060d1a2f"} Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.837708 4948 scope.go:117] "RemoveContainer" containerID="59eceb81df286be225188d13264d44596719c0a06ca1c39903c517b422013a68" Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.842544 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b8ad7a10-6353-45f2-a11d-126d0d4786b0","Type":"ContainerStarted","Data":"b74656e29439723075b3c2514dc8006478ea22c951aec141fc385a6ef5cf24b1"} Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.842661 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="b8ad7a10-6353-45f2-a11d-126d0d4786b0" containerName="glance-log" containerID="cri-o://fdfa9ea96d17b0fbac466add147012945b8f9c77bfe4c7e3939115948c20ff15" gracePeriod=30 Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.842809 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="b8ad7a10-6353-45f2-a11d-126d0d4786b0" containerName="glance-httpd" containerID="cri-o://b74656e29439723075b3c2514dc8006478ea22c951aec141fc385a6ef5cf24b1" gracePeriod=30 Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.855417 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-96668" event={"ID":"94b82d55-b467-408f-8eb1-6da4648dde97","Type":"ContainerStarted","Data":"76e608849a7aba9c1d66b8ca878674048ebd2cf05258476f0e82590946ed89c0"} Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.855573 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.858773 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6e945937-1e9d-465b-a68f-c755119ab6f9","Type":"ContainerStarted","Data":"e38f809977ba5076c8af037b7e38d52d33545f893ce90480e346caeaaef81dee"} Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.906507 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.90648207 podStartE2EDuration="4.90648207s" podCreationTimestamp="2026-02-20 08:22:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:22:18.900492322 +0000 UTC m=+987.874987142" watchObservedRunningTime="2026-02-20 08:22:18.90648207 +0000 UTC m=+987.880976890" Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.916689 4948 scope.go:117] "RemoveContainer" containerID="833390a5ef5c4525790a7cad88740d3ce4d0df639a0829042164aca3b32d8d15" Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.943311 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-96668" podStartSLOduration=4.943287031 podStartE2EDuration="4.943287031s" podCreationTimestamp="2026-02-20 08:22:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:22:18.92670419 +0000 UTC m=+987.901199030" watchObservedRunningTime="2026-02-20 08:22:18.943287031 +0000 UTC m=+987.917781851" Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.973856 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lnp7g"] Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.975517 4948 scope.go:117] "RemoveContainer" containerID="82db4fd040d5ae64aa60b6299d284256aa892181e4cf858ebfad62dc9681817b" Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.979341 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lnp7g"] Feb 20 08:22:19 crc kubenswrapper[4948]: I0220 08:22:19.073130 4948 scope.go:117] "RemoveContainer" containerID="59eceb81df286be225188d13264d44596719c0a06ca1c39903c517b422013a68" Feb 20 08:22:19 crc kubenswrapper[4948]: E0220 08:22:19.075728 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59eceb81df286be225188d13264d44596719c0a06ca1c39903c517b422013a68\": container with ID starting with 59eceb81df286be225188d13264d44596719c0a06ca1c39903c517b422013a68 not found: ID does not exist" containerID="59eceb81df286be225188d13264d44596719c0a06ca1c39903c517b422013a68" Feb 20 08:22:19 crc kubenswrapper[4948]: I0220 08:22:19.075769 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59eceb81df286be225188d13264d44596719c0a06ca1c39903c517b422013a68"} err="failed to get container status \"59eceb81df286be225188d13264d44596719c0a06ca1c39903c517b422013a68\": rpc error: code = NotFound desc = could not find container \"59eceb81df286be225188d13264d44596719c0a06ca1c39903c517b422013a68\": container with ID starting with 59eceb81df286be225188d13264d44596719c0a06ca1c39903c517b422013a68 not found: ID does not exist" Feb 20 08:22:19 crc kubenswrapper[4948]: I0220 08:22:19.075792 4948 scope.go:117] "RemoveContainer" containerID="833390a5ef5c4525790a7cad88740d3ce4d0df639a0829042164aca3b32d8d15" Feb 20 08:22:19 crc kubenswrapper[4948]: E0220 08:22:19.076360 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"833390a5ef5c4525790a7cad88740d3ce4d0df639a0829042164aca3b32d8d15\": container with ID starting with 833390a5ef5c4525790a7cad88740d3ce4d0df639a0829042164aca3b32d8d15 not found: ID does not exist" containerID="833390a5ef5c4525790a7cad88740d3ce4d0df639a0829042164aca3b32d8d15" Feb 20 08:22:19 crc kubenswrapper[4948]: I0220 08:22:19.076381 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"833390a5ef5c4525790a7cad88740d3ce4d0df639a0829042164aca3b32d8d15"} err="failed to get container status \"833390a5ef5c4525790a7cad88740d3ce4d0df639a0829042164aca3b32d8d15\": rpc error: code = NotFound desc = could not find container \"833390a5ef5c4525790a7cad88740d3ce4d0df639a0829042164aca3b32d8d15\": container with ID starting with 833390a5ef5c4525790a7cad88740d3ce4d0df639a0829042164aca3b32d8d15 not found: ID does not exist" Feb 20 08:22:19 crc kubenswrapper[4948]: I0220 08:22:19.076401 4948 scope.go:117] "RemoveContainer" containerID="82db4fd040d5ae64aa60b6299d284256aa892181e4cf858ebfad62dc9681817b" Feb 20 08:22:19 crc kubenswrapper[4948]: E0220 08:22:19.076886 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82db4fd040d5ae64aa60b6299d284256aa892181e4cf858ebfad62dc9681817b\": container with ID starting with 82db4fd040d5ae64aa60b6299d284256aa892181e4cf858ebfad62dc9681817b not found: ID does not exist" containerID="82db4fd040d5ae64aa60b6299d284256aa892181e4cf858ebfad62dc9681817b" Feb 20 08:22:19 crc kubenswrapper[4948]: I0220 08:22:19.076905 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82db4fd040d5ae64aa60b6299d284256aa892181e4cf858ebfad62dc9681817b"} err="failed to get container status \"82db4fd040d5ae64aa60b6299d284256aa892181e4cf858ebfad62dc9681817b\": rpc error: code = NotFound desc = could not find container \"82db4fd040d5ae64aa60b6299d284256aa892181e4cf858ebfad62dc9681817b\": container with ID starting with 82db4fd040d5ae64aa60b6299d284256aa892181e4cf858ebfad62dc9681817b not found: ID does not exist" Feb 20 08:22:19 crc kubenswrapper[4948]: I0220 08:22:19.742076 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81e7d5de-2379-46c7-ba9b-ffcacad79e29" path="/var/lib/kubelet/pods/81e7d5de-2379-46c7-ba9b-ffcacad79e29/volumes" Feb 20 08:22:19 crc kubenswrapper[4948]: I0220 08:22:19.878445 4948 generic.go:334] "Generic (PLEG): container finished" podID="9959c3d4-fb0e-473c-8bbe-065110a3e03a" containerID="fa275de1b67d8651c437fb815c4f8470dc8d77cd9885aaddee1a7419f33fd25c" exitCode=0 Feb 20 08:22:19 crc kubenswrapper[4948]: I0220 08:22:19.878514 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-g72bc" event={"ID":"9959c3d4-fb0e-473c-8bbe-065110a3e03a","Type":"ContainerDied","Data":"fa275de1b67d8651c437fb815c4f8470dc8d77cd9885aaddee1a7419f33fd25c"} Feb 20 08:22:19 crc kubenswrapper[4948]: I0220 08:22:19.885409 4948 generic.go:334] "Generic (PLEG): container finished" podID="b8ad7a10-6353-45f2-a11d-126d0d4786b0" containerID="b74656e29439723075b3c2514dc8006478ea22c951aec141fc385a6ef5cf24b1" exitCode=0 Feb 20 08:22:19 crc kubenswrapper[4948]: I0220 08:22:19.885442 4948 generic.go:334] "Generic (PLEG): container finished" podID="b8ad7a10-6353-45f2-a11d-126d0d4786b0" containerID="fdfa9ea96d17b0fbac466add147012945b8f9c77bfe4c7e3939115948c20ff15" exitCode=143 Feb 20 08:22:19 crc kubenswrapper[4948]: I0220 08:22:19.885513 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b8ad7a10-6353-45f2-a11d-126d0d4786b0","Type":"ContainerDied","Data":"b74656e29439723075b3c2514dc8006478ea22c951aec141fc385a6ef5cf24b1"} Feb 20 08:22:19 crc kubenswrapper[4948]: I0220 08:22:19.885542 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b8ad7a10-6353-45f2-a11d-126d0d4786b0","Type":"ContainerDied","Data":"fdfa9ea96d17b0fbac466add147012945b8f9c77bfe4c7e3939115948c20ff15"} Feb 20 08:22:19 crc kubenswrapper[4948]: I0220 08:22:19.889623 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6e945937-1e9d-465b-a68f-c755119ab6f9","Type":"ContainerStarted","Data":"344f41da66c91718be93d872d218e57c352e04b55c97d3dc69d71b942dbe58d1"} Feb 20 08:22:19 crc kubenswrapper[4948]: I0220 08:22:19.890037 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="6e945937-1e9d-465b-a68f-c755119ab6f9" containerName="glance-log" containerID="cri-o://e38f809977ba5076c8af037b7e38d52d33545f893ce90480e346caeaaef81dee" gracePeriod=30 Feb 20 08:22:19 crc kubenswrapper[4948]: I0220 08:22:19.890185 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="6e945937-1e9d-465b-a68f-c755119ab6f9" containerName="glance-httpd" containerID="cri-o://344f41da66c91718be93d872d218e57c352e04b55c97d3dc69d71b942dbe58d1" gracePeriod=30 Feb 20 08:22:19 crc kubenswrapper[4948]: I0220 08:22:19.926328 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.926307868 podStartE2EDuration="5.926307868s" podCreationTimestamp="2026-02-20 08:22:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:22:19.920483574 +0000 UTC m=+988.894978394" watchObservedRunningTime="2026-02-20 08:22:19.926307868 +0000 UTC m=+988.900802688" Feb 20 08:22:20 crc kubenswrapper[4948]: I0220 08:22:20.904558 4948 generic.go:334] "Generic (PLEG): container finished" podID="6e945937-1e9d-465b-a68f-c755119ab6f9" containerID="344f41da66c91718be93d872d218e57c352e04b55c97d3dc69d71b942dbe58d1" exitCode=0 Feb 20 08:22:20 crc kubenswrapper[4948]: I0220 08:22:20.905557 4948 generic.go:334] "Generic (PLEG): container finished" podID="6e945937-1e9d-465b-a68f-c755119ab6f9" containerID="e38f809977ba5076c8af037b7e38d52d33545f893ce90480e346caeaaef81dee" exitCode=143 Feb 20 08:22:20 crc kubenswrapper[4948]: I0220 08:22:20.904904 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6e945937-1e9d-465b-a68f-c755119ab6f9","Type":"ContainerDied","Data":"344f41da66c91718be93d872d218e57c352e04b55c97d3dc69d71b942dbe58d1"} Feb 20 08:22:20 crc kubenswrapper[4948]: I0220 08:22:20.905814 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6e945937-1e9d-465b-a68f-c755119ab6f9","Type":"ContainerDied","Data":"e38f809977ba5076c8af037b7e38d52d33545f893ce90480e346caeaaef81dee"} Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.347668 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.452475 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-scripts\") pod \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.452564 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqhps\" (UniqueName: \"kubernetes.io/projected/b8ad7a10-6353-45f2-a11d-126d0d4786b0-kube-api-access-kqhps\") pod \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.452606 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8ad7a10-6353-45f2-a11d-126d0d4786b0-logs\") pod \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.452678 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b8ad7a10-6353-45f2-a11d-126d0d4786b0-httpd-run\") pod \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.452842 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-combined-ca-bundle\") pod \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.452893 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.452950 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-config-data\") pod \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.453043 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8ad7a10-6353-45f2-a11d-126d0d4786b0-logs" (OuterVolumeSpecName: "logs") pod "b8ad7a10-6353-45f2-a11d-126d0d4786b0" (UID: "b8ad7a10-6353-45f2-a11d-126d0d4786b0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.453115 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8ad7a10-6353-45f2-a11d-126d0d4786b0-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b8ad7a10-6353-45f2-a11d-126d0d4786b0" (UID: "b8ad7a10-6353-45f2-a11d-126d0d4786b0"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.453398 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8ad7a10-6353-45f2-a11d-126d0d4786b0-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.453420 4948 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b8ad7a10-6353-45f2-a11d-126d0d4786b0-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.459662 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8ad7a10-6353-45f2-a11d-126d0d4786b0-kube-api-access-kqhps" (OuterVolumeSpecName: "kube-api-access-kqhps") pod "b8ad7a10-6353-45f2-a11d-126d0d4786b0" (UID: "b8ad7a10-6353-45f2-a11d-126d0d4786b0"). InnerVolumeSpecName "kube-api-access-kqhps". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.460125 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-scripts" (OuterVolumeSpecName: "scripts") pod "b8ad7a10-6353-45f2-a11d-126d0d4786b0" (UID: "b8ad7a10-6353-45f2-a11d-126d0d4786b0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.460445 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "b8ad7a10-6353-45f2-a11d-126d0d4786b0" (UID: "b8ad7a10-6353-45f2-a11d-126d0d4786b0"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.499564 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b8ad7a10-6353-45f2-a11d-126d0d4786b0" (UID: "b8ad7a10-6353-45f2-a11d-126d0d4786b0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.520145 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-config-data" (OuterVolumeSpecName: "config-data") pod "b8ad7a10-6353-45f2-a11d-126d0d4786b0" (UID: "b8ad7a10-6353-45f2-a11d-126d0d4786b0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.555201 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqhps\" (UniqueName: \"kubernetes.io/projected/b8ad7a10-6353-45f2-a11d-126d0d4786b0-kube-api-access-kqhps\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.555246 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.555288 4948 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.555304 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.555317 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.577402 4948 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.658405 4948 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.937212 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b8ad7a10-6353-45f2-a11d-126d0d4786b0","Type":"ContainerDied","Data":"434c4fc6de7f62028433ff88fb8c531a01a506655fd56b116472009d2d35dd9f"} Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.937296 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.937382 4948 scope.go:117] "RemoveContainer" containerID="b74656e29439723075b3c2514dc8006478ea22c951aec141fc385a6ef5cf24b1" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.970516 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.978284 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.997840 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:22:22 crc kubenswrapper[4948]: E0220 08:22:22.998361 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69c38b92-a051-4834-bf34-8612edb31dfa" containerName="init" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.998380 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="69c38b92-a051-4834-bf34-8612edb31dfa" containerName="init" Feb 20 08:22:22 crc kubenswrapper[4948]: E0220 08:22:22.998415 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8ad7a10-6353-45f2-a11d-126d0d4786b0" containerName="glance-log" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.998429 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8ad7a10-6353-45f2-a11d-126d0d4786b0" containerName="glance-log" Feb 20 08:22:22 crc kubenswrapper[4948]: E0220 08:22:22.998444 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7960b704-21ac-4d75-8226-a9bbae4c2b50" containerName="init" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.998450 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="7960b704-21ac-4d75-8226-a9bbae4c2b50" containerName="init" Feb 20 08:22:22 crc kubenswrapper[4948]: E0220 08:22:22.998458 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81e7d5de-2379-46c7-ba9b-ffcacad79e29" containerName="registry-server" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.998464 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="81e7d5de-2379-46c7-ba9b-ffcacad79e29" containerName="registry-server" Feb 20 08:22:22 crc kubenswrapper[4948]: E0220 08:22:22.998511 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8ad7a10-6353-45f2-a11d-126d0d4786b0" containerName="glance-httpd" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.998519 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8ad7a10-6353-45f2-a11d-126d0d4786b0" containerName="glance-httpd" Feb 20 08:22:22 crc kubenswrapper[4948]: E0220 08:22:22.998530 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81e7d5de-2379-46c7-ba9b-ffcacad79e29" containerName="extract-utilities" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.998537 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="81e7d5de-2379-46c7-ba9b-ffcacad79e29" containerName="extract-utilities" Feb 20 08:22:22 crc kubenswrapper[4948]: E0220 08:22:22.998553 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81e7d5de-2379-46c7-ba9b-ffcacad79e29" containerName="extract-content" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.998560 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="81e7d5de-2379-46c7-ba9b-ffcacad79e29" containerName="extract-content" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.998770 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8ad7a10-6353-45f2-a11d-126d0d4786b0" containerName="glance-log" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.998789 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="81e7d5de-2379-46c7-ba9b-ffcacad79e29" containerName="registry-server" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.998801 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="69c38b92-a051-4834-bf34-8612edb31dfa" containerName="init" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.998836 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="7960b704-21ac-4d75-8226-a9bbae4c2b50" containerName="init" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.998845 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8ad7a10-6353-45f2-a11d-126d0d4786b0" containerName="glance-httpd" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.999961 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.002146 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.016517 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.066617 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5zcb\" (UniqueName: \"kubernetes.io/projected/e0f82f51-8a20-402d-8861-6c57a4e5af0d-kube-api-access-q5zcb\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.066875 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0f82f51-8a20-402d-8861-6c57a4e5af0d-logs\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.066996 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e0f82f51-8a20-402d-8861-6c57a4e5af0d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.067094 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-config-data\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.067205 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-scripts\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.067318 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.067420 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.169005 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5zcb\" (UniqueName: \"kubernetes.io/projected/e0f82f51-8a20-402d-8861-6c57a4e5af0d-kube-api-access-q5zcb\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.169399 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0f82f51-8a20-402d-8861-6c57a4e5af0d-logs\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.169434 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e0f82f51-8a20-402d-8861-6c57a4e5af0d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.169458 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-config-data\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.169492 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-scripts\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.169524 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.169560 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.170186 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.170808 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e0f82f51-8a20-402d-8861-6c57a4e5af0d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.171698 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0f82f51-8a20-402d-8861-6c57a4e5af0d-logs\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.174712 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-scripts\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.176528 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.176868 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-config-data\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.196602 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5zcb\" (UniqueName: \"kubernetes.io/projected/e0f82f51-8a20-402d-8861-6c57a4e5af0d-kube-api-access-q5zcb\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.209697 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.328626 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.735647 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8ad7a10-6353-45f2-a11d-126d0d4786b0" path="/var/lib/kubelet/pods/b8ad7a10-6353-45f2-a11d-126d0d4786b0/volumes" Feb 20 08:22:24 crc kubenswrapper[4948]: I0220 08:22:24.712376 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:22:24 crc kubenswrapper[4948]: I0220 08:22:24.970196 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.045053 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-tkqzb"] Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.046812 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-tkqzb" podUID="da816a64-160f-4a5c-b945-b8fa467208d6" containerName="dnsmasq-dns" containerID="cri-o://562bf68215847fe83f99073d009b1ee89904610f3d702b2ab282baecbca78fb2" gracePeriod=10 Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.074775 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-54d5f75fb9-s2659"] Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.125462 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-66b4d4cb8-hgmkd"] Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.129906 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.138018 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.156883 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-66b4d4cb8-hgmkd"] Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.228639 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-68c9c4f5fc-2vtb4"] Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.300285 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-685c4db87b-rw57h"] Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.301784 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.305913 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d1046503-c6d3-49bf-9523-2f897fcb82ba-scripts\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.306009 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-horizon-secret-key\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.306040 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d1046503-c6d3-49bf-9523-2f897fcb82ba-config-data\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.306060 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-horizon-tls-certs\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.306084 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgwfv\" (UniqueName: \"kubernetes.io/projected/d1046503-c6d3-49bf-9523-2f897fcb82ba-kube-api-access-mgwfv\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.306125 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d1046503-c6d3-49bf-9523-2f897fcb82ba-logs\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.306175 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-combined-ca-bundle\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.320719 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-685c4db87b-rw57h"] Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.407891 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7b9a072a-a040-414c-96c7-cf1454148745-scripts\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.408204 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7b9a072a-a040-414c-96c7-cf1454148745-config-data\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.408299 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d1046503-c6d3-49bf-9523-2f897fcb82ba-scripts\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.408374 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b9a072a-a040-414c-96c7-cf1454148745-combined-ca-bundle\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.408536 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-horizon-secret-key\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.408585 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d1046503-c6d3-49bf-9523-2f897fcb82ba-config-data\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.408605 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-horizon-tls-certs\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.408629 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgwfv\" (UniqueName: \"kubernetes.io/projected/d1046503-c6d3-49bf-9523-2f897fcb82ba-kube-api-access-mgwfv\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.408672 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdmtj\" (UniqueName: \"kubernetes.io/projected/7b9a072a-a040-414c-96c7-cf1454148745-kube-api-access-vdmtj\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.408706 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b9a072a-a040-414c-96c7-cf1454148745-horizon-tls-certs\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.408729 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d1046503-c6d3-49bf-9523-2f897fcb82ba-logs\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.408830 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b9a072a-a040-414c-96c7-cf1454148745-logs\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.408867 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-combined-ca-bundle\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.408930 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7b9a072a-a040-414c-96c7-cf1454148745-horizon-secret-key\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.409466 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d1046503-c6d3-49bf-9523-2f897fcb82ba-scripts\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.413831 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d1046503-c6d3-49bf-9523-2f897fcb82ba-config-data\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.419843 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d1046503-c6d3-49bf-9523-2f897fcb82ba-logs\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.421481 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-horizon-secret-key\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.421941 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-combined-ca-bundle\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.422610 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-horizon-tls-certs\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.449528 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgwfv\" (UniqueName: \"kubernetes.io/projected/d1046503-c6d3-49bf-9523-2f897fcb82ba-kube-api-access-mgwfv\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.476606 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.510445 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7b9a072a-a040-414c-96c7-cf1454148745-config-data\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.510495 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b9a072a-a040-414c-96c7-cf1454148745-combined-ca-bundle\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.510540 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdmtj\" (UniqueName: \"kubernetes.io/projected/7b9a072a-a040-414c-96c7-cf1454148745-kube-api-access-vdmtj\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.510562 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b9a072a-a040-414c-96c7-cf1454148745-horizon-tls-certs\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.510591 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b9a072a-a040-414c-96c7-cf1454148745-logs\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.510626 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7b9a072a-a040-414c-96c7-cf1454148745-horizon-secret-key\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.510676 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7b9a072a-a040-414c-96c7-cf1454148745-scripts\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.511363 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7b9a072a-a040-414c-96c7-cf1454148745-scripts\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.512201 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7b9a072a-a040-414c-96c7-cf1454148745-config-data\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.514152 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b9a072a-a040-414c-96c7-cf1454148745-logs\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.518712 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b9a072a-a040-414c-96c7-cf1454148745-horizon-tls-certs\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.519441 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b9a072a-a040-414c-96c7-cf1454148745-combined-ca-bundle\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.534454 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7b9a072a-a040-414c-96c7-cf1454148745-horizon-secret-key\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.542712 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdmtj\" (UniqueName: \"kubernetes.io/projected/7b9a072a-a040-414c-96c7-cf1454148745-kube-api-access-vdmtj\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.626222 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-g27hk"] Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.627789 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g27hk" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.637281 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g27hk"] Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.642339 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.815515 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6729921-0a12-46a0-824f-810df88ae426-catalog-content\") pod \"community-operators-g27hk\" (UID: \"d6729921-0a12-46a0-824f-810df88ae426\") " pod="openshift-marketplace/community-operators-g27hk" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.815599 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2k5c\" (UniqueName: \"kubernetes.io/projected/d6729921-0a12-46a0-824f-810df88ae426-kube-api-access-x2k5c\") pod \"community-operators-g27hk\" (UID: \"d6729921-0a12-46a0-824f-810df88ae426\") " pod="openshift-marketplace/community-operators-g27hk" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.815991 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6729921-0a12-46a0-824f-810df88ae426-utilities\") pod \"community-operators-g27hk\" (UID: \"d6729921-0a12-46a0-824f-810df88ae426\") " pod="openshift-marketplace/community-operators-g27hk" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.918169 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6729921-0a12-46a0-824f-810df88ae426-utilities\") pod \"community-operators-g27hk\" (UID: \"d6729921-0a12-46a0-824f-810df88ae426\") " pod="openshift-marketplace/community-operators-g27hk" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.918238 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6729921-0a12-46a0-824f-810df88ae426-catalog-content\") pod \"community-operators-g27hk\" (UID: \"d6729921-0a12-46a0-824f-810df88ae426\") " pod="openshift-marketplace/community-operators-g27hk" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.918302 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2k5c\" (UniqueName: \"kubernetes.io/projected/d6729921-0a12-46a0-824f-810df88ae426-kube-api-access-x2k5c\") pod \"community-operators-g27hk\" (UID: \"d6729921-0a12-46a0-824f-810df88ae426\") " pod="openshift-marketplace/community-operators-g27hk" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.918759 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6729921-0a12-46a0-824f-810df88ae426-utilities\") pod \"community-operators-g27hk\" (UID: \"d6729921-0a12-46a0-824f-810df88ae426\") " pod="openshift-marketplace/community-operators-g27hk" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.918815 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6729921-0a12-46a0-824f-810df88ae426-catalog-content\") pod \"community-operators-g27hk\" (UID: \"d6729921-0a12-46a0-824f-810df88ae426\") " pod="openshift-marketplace/community-operators-g27hk" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.937550 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2k5c\" (UniqueName: \"kubernetes.io/projected/d6729921-0a12-46a0-824f-810df88ae426-kube-api-access-x2k5c\") pod \"community-operators-g27hk\" (UID: \"d6729921-0a12-46a0-824f-810df88ae426\") " pod="openshift-marketplace/community-operators-g27hk" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.965398 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g27hk" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.974359 4948 generic.go:334] "Generic (PLEG): container finished" podID="da816a64-160f-4a5c-b945-b8fa467208d6" containerID="562bf68215847fe83f99073d009b1ee89904610f3d702b2ab282baecbca78fb2" exitCode=0 Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.974404 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-tkqzb" event={"ID":"da816a64-160f-4a5c-b945-b8fa467208d6","Type":"ContainerDied","Data":"562bf68215847fe83f99073d009b1ee89904610f3d702b2ab282baecbca78fb2"} Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.349965 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.543773 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-combined-ca-bundle\") pod \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.543856 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p27c6\" (UniqueName: \"kubernetes.io/projected/9959c3d4-fb0e-473c-8bbe-065110a3e03a-kube-api-access-p27c6\") pod \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.543926 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-credential-keys\") pod \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.543964 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-scripts\") pod \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.544009 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-config-data\") pod \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.544033 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-fernet-keys\") pod \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.549731 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "9959c3d4-fb0e-473c-8bbe-065110a3e03a" (UID: "9959c3d4-fb0e-473c-8bbe-065110a3e03a"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.549759 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-scripts" (OuterVolumeSpecName: "scripts") pod "9959c3d4-fb0e-473c-8bbe-065110a3e03a" (UID: "9959c3d4-fb0e-473c-8bbe-065110a3e03a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.550273 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9959c3d4-fb0e-473c-8bbe-065110a3e03a-kube-api-access-p27c6" (OuterVolumeSpecName: "kube-api-access-p27c6") pod "9959c3d4-fb0e-473c-8bbe-065110a3e03a" (UID: "9959c3d4-fb0e-473c-8bbe-065110a3e03a"). InnerVolumeSpecName "kube-api-access-p27c6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.550554 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "9959c3d4-fb0e-473c-8bbe-065110a3e03a" (UID: "9959c3d4-fb0e-473c-8bbe-065110a3e03a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.569079 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9959c3d4-fb0e-473c-8bbe-065110a3e03a" (UID: "9959c3d4-fb0e-473c-8bbe-065110a3e03a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.588870 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-config-data" (OuterVolumeSpecName: "config-data") pod "9959c3d4-fb0e-473c-8bbe-065110a3e03a" (UID: "9959c3d4-fb0e-473c-8bbe-065110a3e03a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.645685 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.646044 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p27c6\" (UniqueName: \"kubernetes.io/projected/9959c3d4-fb0e-473c-8bbe-065110a3e03a-kube-api-access-p27c6\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.646062 4948 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.646074 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.646086 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.646096 4948 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.994828 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-g72bc" event={"ID":"9959c3d4-fb0e-473c-8bbe-065110a3e03a","Type":"ContainerDied","Data":"1eedce1f0a19f6576c482c3583bf052a9786f8dcd27e5bce2d34fd0ef16f70ad"} Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.994876 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1eedce1f0a19f6576c482c3583bf052a9786f8dcd27e5bce2d34fd0ef16f70ad" Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.994954 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.434314 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-g72bc"] Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.442021 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-g72bc"] Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.535304 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-v29fx"] Feb 20 08:22:28 crc kubenswrapper[4948]: E0220 08:22:28.535935 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9959c3d4-fb0e-473c-8bbe-065110a3e03a" containerName="keystone-bootstrap" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.536062 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="9959c3d4-fb0e-473c-8bbe-065110a3e03a" containerName="keystone-bootstrap" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.536387 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="9959c3d4-fb0e-473c-8bbe-065110a3e03a" containerName="keystone-bootstrap" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.537310 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.540273 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.540551 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-v9rj2" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.540684 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.541101 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.541623 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.545387 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-v29fx"] Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.581822 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-scripts\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.581871 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-credential-keys\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.581907 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-config-data\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.581949 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-fernet-keys\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.582068 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-combined-ca-bundle\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.582117 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jr9d\" (UniqueName: \"kubernetes.io/projected/d84a71b6-7503-4150-a5c4-4579b08a669a-kube-api-access-6jr9d\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.684153 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-config-data\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.684243 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-fernet-keys\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.684280 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-combined-ca-bundle\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.684339 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jr9d\" (UniqueName: \"kubernetes.io/projected/d84a71b6-7503-4150-a5c4-4579b08a669a-kube-api-access-6jr9d\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.684428 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-scripts\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.684469 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-credential-keys\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.691760 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-config-data\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.696453 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-fernet-keys\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.701517 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-scripts\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.702041 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-combined-ca-bundle\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.703728 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-credential-keys\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.706380 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jr9d\" (UniqueName: \"kubernetes.io/projected/d84a71b6-7503-4150-a5c4-4579b08a669a-kube-api-access-6jr9d\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.866921 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:29 crc kubenswrapper[4948]: I0220 08:22:28.999704 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hh5s2"] Feb 20 08:22:29 crc kubenswrapper[4948]: I0220 08:22:29.002917 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hh5s2" Feb 20 08:22:29 crc kubenswrapper[4948]: I0220 08:22:29.015136 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hh5s2"] Feb 20 08:22:29 crc kubenswrapper[4948]: I0220 08:22:29.095213 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/426b37f9-4acb-4d28-a55c-3fcde2600002-utilities\") pod \"certified-operators-hh5s2\" (UID: \"426b37f9-4acb-4d28-a55c-3fcde2600002\") " pod="openshift-marketplace/certified-operators-hh5s2" Feb 20 08:22:29 crc kubenswrapper[4948]: I0220 08:22:29.095548 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tw6bb\" (UniqueName: \"kubernetes.io/projected/426b37f9-4acb-4d28-a55c-3fcde2600002-kube-api-access-tw6bb\") pod \"certified-operators-hh5s2\" (UID: \"426b37f9-4acb-4d28-a55c-3fcde2600002\") " pod="openshift-marketplace/certified-operators-hh5s2" Feb 20 08:22:29 crc kubenswrapper[4948]: I0220 08:22:29.095702 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/426b37f9-4acb-4d28-a55c-3fcde2600002-catalog-content\") pod \"certified-operators-hh5s2\" (UID: \"426b37f9-4acb-4d28-a55c-3fcde2600002\") " pod="openshift-marketplace/certified-operators-hh5s2" Feb 20 08:22:29 crc kubenswrapper[4948]: I0220 08:22:29.199741 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tw6bb\" (UniqueName: \"kubernetes.io/projected/426b37f9-4acb-4d28-a55c-3fcde2600002-kube-api-access-tw6bb\") pod \"certified-operators-hh5s2\" (UID: \"426b37f9-4acb-4d28-a55c-3fcde2600002\") " pod="openshift-marketplace/certified-operators-hh5s2" Feb 20 08:22:29 crc kubenswrapper[4948]: I0220 08:22:29.199809 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/426b37f9-4acb-4d28-a55c-3fcde2600002-catalog-content\") pod \"certified-operators-hh5s2\" (UID: \"426b37f9-4acb-4d28-a55c-3fcde2600002\") " pod="openshift-marketplace/certified-operators-hh5s2" Feb 20 08:22:29 crc kubenswrapper[4948]: I0220 08:22:29.199873 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/426b37f9-4acb-4d28-a55c-3fcde2600002-utilities\") pod \"certified-operators-hh5s2\" (UID: \"426b37f9-4acb-4d28-a55c-3fcde2600002\") " pod="openshift-marketplace/certified-operators-hh5s2" Feb 20 08:22:29 crc kubenswrapper[4948]: I0220 08:22:29.200369 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/426b37f9-4acb-4d28-a55c-3fcde2600002-utilities\") pod \"certified-operators-hh5s2\" (UID: \"426b37f9-4acb-4d28-a55c-3fcde2600002\") " pod="openshift-marketplace/certified-operators-hh5s2" Feb 20 08:22:29 crc kubenswrapper[4948]: I0220 08:22:29.200893 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/426b37f9-4acb-4d28-a55c-3fcde2600002-catalog-content\") pod \"certified-operators-hh5s2\" (UID: \"426b37f9-4acb-4d28-a55c-3fcde2600002\") " pod="openshift-marketplace/certified-operators-hh5s2" Feb 20 08:22:29 crc kubenswrapper[4948]: I0220 08:22:29.220886 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tw6bb\" (UniqueName: \"kubernetes.io/projected/426b37f9-4acb-4d28-a55c-3fcde2600002-kube-api-access-tw6bb\") pod \"certified-operators-hh5s2\" (UID: \"426b37f9-4acb-4d28-a55c-3fcde2600002\") " pod="openshift-marketplace/certified-operators-hh5s2" Feb 20 08:22:29 crc kubenswrapper[4948]: I0220 08:22:29.358061 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hh5s2" Feb 20 08:22:29 crc kubenswrapper[4948]: I0220 08:22:29.732132 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9959c3d4-fb0e-473c-8bbe-065110a3e03a" path="/var/lib/kubelet/pods/9959c3d4-fb0e-473c-8bbe-065110a3e03a/volumes" Feb 20 08:22:29 crc kubenswrapper[4948]: I0220 08:22:29.848630 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-tkqzb" podUID="da816a64-160f-4a5c-b945-b8fa467208d6" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: connect: connection refused" Feb 20 08:22:32 crc kubenswrapper[4948]: E0220 08:22:32.527602 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Feb 20 08:22:32 crc kubenswrapper[4948]: E0220 08:22:32.529307 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nb6h656h698h57bh5dbhcbh5bh567h57dh56ch99hb4h68fh58h597h9fh5d5h55fh66ch554h57dhbdh67bh8dh56ch567h8chb8h58fh74h5c6h656q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vp59z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-68c9c4f5fc-2vtb4_openstack(136dd621-a5c5-481a-9b86-66f8cb72486d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 08:22:32 crc kubenswrapper[4948]: E0220 08:22:32.532557 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-68c9c4f5fc-2vtb4" podUID="136dd621-a5c5-481a-9b86-66f8cb72486d" Feb 20 08:22:32 crc kubenswrapper[4948]: E0220 08:22:32.559886 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Feb 20 08:22:32 crc kubenswrapper[4948]: E0220 08:22:32.560446 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nc9h65bh5fh9bhbch96h5cch548h5fdh57bh5c8h5b5h667h64h74h566h558h54bh64fh9fhd6h694h589h5c6h5d5h4hbfh577hdh5d4h7fhfcq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gxqmm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-54d5f75fb9-s2659_openstack(549beb55-bab4-4b1a-bf98-48de2e12956f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 08:22:32 crc kubenswrapper[4948]: E0220 08:22:32.565917 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-54d5f75fb9-s2659" podUID="549beb55-bab4-4b1a-bf98-48de2e12956f" Feb 20 08:22:32 crc kubenswrapper[4948]: E0220 08:22:32.567298 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Feb 20 08:22:32 crc kubenswrapper[4948]: E0220 08:22:32.567455 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n89h589h676hfch59ch5c8h699h586h5bfh74hd5h54h599h586h56fh99h54dhch8fh668h678h5b6h694hcfh8ch564h654h66dh5dbh554h5c9h647q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jrsxt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7d578c747f-8f6hc_openstack(8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 08:22:32 crc kubenswrapper[4948]: E0220 08:22:32.570333 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7d578c747f-8f6hc" podUID="8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4" Feb 20 08:22:36 crc kubenswrapper[4948]: I0220 08:22:36.065516 4948 generic.go:334] "Generic (PLEG): container finished" podID="bb362d0e-68b4-4d48-885c-a6894ac71e6e" containerID="cb7241154041475439806eb631212532adb6e6775adf70e62bdf5cf4b6d19276" exitCode=0 Feb 20 08:22:36 crc kubenswrapper[4948]: I0220 08:22:36.065680 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-g5zxh" event={"ID":"bb362d0e-68b4-4d48-885c-a6894ac71e6e","Type":"ContainerDied","Data":"cb7241154041475439806eb631212532adb6e6775adf70e62bdf5cf4b6d19276"} Feb 20 08:22:39 crc kubenswrapper[4948]: I0220 08:22:39.849489 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-tkqzb" podUID="da816a64-160f-4a5c-b945-b8fa467208d6" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: i/o timeout" Feb 20 08:22:40 crc kubenswrapper[4948]: E0220 08:22:40.010363 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Feb 20 08:22:40 crc kubenswrapper[4948]: E0220 08:22:40.010575 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gmlgz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-c7zzv_openstack(f1f108e1-27bb-433e-b7e1-f9cc7f778182): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 08:22:40 crc kubenswrapper[4948]: E0220 08:22:40.011719 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-c7zzv" podUID="f1f108e1-27bb-433e-b7e1-f9cc7f778182" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.079797 4948 scope.go:117] "RemoveContainer" containerID="fdfa9ea96d17b0fbac466add147012945b8f9c77bfe4c7e3939115948c20ff15" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.102156 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-tkqzb" event={"ID":"da816a64-160f-4a5c-b945-b8fa467208d6","Type":"ContainerDied","Data":"32b2c84d98711b304188499ff3c33cd5bf7af6d44c7a701d021f0522c169184b"} Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.102421 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32b2c84d98711b304188499ff3c33cd5bf7af6d44c7a701d021f0522c169184b" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.110075 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6e945937-1e9d-465b-a68f-c755119ab6f9","Type":"ContainerDied","Data":"feaec65599576490d6ae435b0c4ffac29904876fc82b7f96fd44c9df4357c090"} Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.110149 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="feaec65599576490d6ae435b0c4ffac29904876fc82b7f96fd44c9df4357c090" Feb 20 08:22:40 crc kubenswrapper[4948]: E0220 08:22:40.111209 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-c7zzv" podUID="f1f108e1-27bb-433e-b7e1-f9cc7f778182" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.225916 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.232393 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.237478 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.246885 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.262959 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.278197 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-g5zxh" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.408854 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-logs\") pod \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.408922 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vp59z\" (UniqueName: \"kubernetes.io/projected/136dd621-a5c5-481a-9b86-66f8cb72486d-kube-api-access-vp59z\") pod \"136dd621-a5c5-481a-9b86-66f8cb72486d\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409115 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/549beb55-bab4-4b1a-bf98-48de2e12956f-logs\") pod \"549beb55-bab4-4b1a-bf98-48de2e12956f\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409153 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svj9w\" (UniqueName: \"kubernetes.io/projected/6e945937-1e9d-465b-a68f-c755119ab6f9-kube-api-access-svj9w\") pod \"6e945937-1e9d-465b-a68f-c755119ab6f9\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409197 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wc2m\" (UniqueName: \"kubernetes.io/projected/da816a64-160f-4a5c-b945-b8fa467208d6-kube-api-access-4wc2m\") pod \"da816a64-160f-4a5c-b945-b8fa467208d6\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409257 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-scripts\") pod \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409312 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-scripts\") pod \"6e945937-1e9d-465b-a68f-c755119ab6f9\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409350 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-config-data\") pod \"6e945937-1e9d-465b-a68f-c755119ab6f9\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409382 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/136dd621-a5c5-481a-9b86-66f8cb72486d-config-data\") pod \"136dd621-a5c5-481a-9b86-66f8cb72486d\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409419 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-combined-ca-bundle\") pod \"6e945937-1e9d-465b-a68f-c755119ab6f9\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409462 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/549beb55-bab4-4b1a-bf98-48de2e12956f-config-data\") pod \"549beb55-bab4-4b1a-bf98-48de2e12956f\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409496 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-config-data\") pod \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409549 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-dns-svc\") pod \"da816a64-160f-4a5c-b945-b8fa467208d6\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409584 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npm75\" (UniqueName: \"kubernetes.io/projected/bb362d0e-68b4-4d48-885c-a6894ac71e6e-kube-api-access-npm75\") pod \"bb362d0e-68b4-4d48-885c-a6894ac71e6e\" (UID: \"bb362d0e-68b4-4d48-885c-a6894ac71e6e\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409637 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-horizon-secret-key\") pod \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409675 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb362d0e-68b4-4d48-885c-a6894ac71e6e-combined-ca-bundle\") pod \"bb362d0e-68b4-4d48-885c-a6894ac71e6e\" (UID: \"bb362d0e-68b4-4d48-885c-a6894ac71e6e\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409723 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/136dd621-a5c5-481a-9b86-66f8cb72486d-logs\") pod \"136dd621-a5c5-481a-9b86-66f8cb72486d\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409774 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrsxt\" (UniqueName: \"kubernetes.io/projected/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-kube-api-access-jrsxt\") pod \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409805 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxqmm\" (UniqueName: \"kubernetes.io/projected/549beb55-bab4-4b1a-bf98-48de2e12956f-kube-api-access-gxqmm\") pod \"549beb55-bab4-4b1a-bf98-48de2e12956f\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409846 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6e945937-1e9d-465b-a68f-c755119ab6f9-httpd-run\") pod \"6e945937-1e9d-465b-a68f-c755119ab6f9\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409878 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-ovsdbserver-sb\") pod \"da816a64-160f-4a5c-b945-b8fa467208d6\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409928 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-config\") pod \"da816a64-160f-4a5c-b945-b8fa467208d6\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409994 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-ovsdbserver-nb\") pod \"da816a64-160f-4a5c-b945-b8fa467208d6\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.410035 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bb362d0e-68b4-4d48-885c-a6894ac71e6e-config\") pod \"bb362d0e-68b4-4d48-885c-a6894ac71e6e\" (UID: \"bb362d0e-68b4-4d48-885c-a6894ac71e6e\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.410065 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e945937-1e9d-465b-a68f-c755119ab6f9-logs\") pod \"6e945937-1e9d-465b-a68f-c755119ab6f9\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.410116 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/549beb55-bab4-4b1a-bf98-48de2e12956f-scripts\") pod \"549beb55-bab4-4b1a-bf98-48de2e12956f\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.410156 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/136dd621-a5c5-481a-9b86-66f8cb72486d-scripts\") pod \"136dd621-a5c5-481a-9b86-66f8cb72486d\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.410187 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/549beb55-bab4-4b1a-bf98-48de2e12956f-horizon-secret-key\") pod \"549beb55-bab4-4b1a-bf98-48de2e12956f\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.410233 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/136dd621-a5c5-481a-9b86-66f8cb72486d-horizon-secret-key\") pod \"136dd621-a5c5-481a-9b86-66f8cb72486d\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.410258 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"6e945937-1e9d-465b-a68f-c755119ab6f9\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.411213 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-logs" (OuterVolumeSpecName: "logs") pod "8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4" (UID: "8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.411865 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/549beb55-bab4-4b1a-bf98-48de2e12956f-scripts" (OuterVolumeSpecName: "scripts") pod "549beb55-bab4-4b1a-bf98-48de2e12956f" (UID: "549beb55-bab4-4b1a-bf98-48de2e12956f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.411962 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/549beb55-bab4-4b1a-bf98-48de2e12956f-config-data" (OuterVolumeSpecName: "config-data") pod "549beb55-bab4-4b1a-bf98-48de2e12956f" (UID: "549beb55-bab4-4b1a-bf98-48de2e12956f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.412505 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-config-data" (OuterVolumeSpecName: "config-data") pod "8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4" (UID: "8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.413145 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/549beb55-bab4-4b1a-bf98-48de2e12956f-logs" (OuterVolumeSpecName: "logs") pod "549beb55-bab4-4b1a-bf98-48de2e12956f" (UID: "549beb55-bab4-4b1a-bf98-48de2e12956f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.412812 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e945937-1e9d-465b-a68f-c755119ab6f9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "6e945937-1e9d-465b-a68f-c755119ab6f9" (UID: "6e945937-1e9d-465b-a68f-c755119ab6f9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.413523 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/136dd621-a5c5-481a-9b86-66f8cb72486d-logs" (OuterVolumeSpecName: "logs") pod "136dd621-a5c5-481a-9b86-66f8cb72486d" (UID: "136dd621-a5c5-481a-9b86-66f8cb72486d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.413737 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e945937-1e9d-465b-a68f-c755119ab6f9-logs" (OuterVolumeSpecName: "logs") pod "6e945937-1e9d-465b-a68f-c755119ab6f9" (UID: "6e945937-1e9d-465b-a68f-c755119ab6f9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.413735 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/136dd621-a5c5-481a-9b86-66f8cb72486d-config-data" (OuterVolumeSpecName: "config-data") pod "136dd621-a5c5-481a-9b86-66f8cb72486d" (UID: "136dd621-a5c5-481a-9b86-66f8cb72486d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.413765 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-scripts" (OuterVolumeSpecName: "scripts") pod "8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4" (UID: "8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.413940 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/136dd621-a5c5-481a-9b86-66f8cb72486d-scripts" (OuterVolumeSpecName: "scripts") pod "136dd621-a5c5-481a-9b86-66f8cb72486d" (UID: "136dd621-a5c5-481a-9b86-66f8cb72486d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.419538 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/136dd621-a5c5-481a-9b86-66f8cb72486d-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "136dd621-a5c5-481a-9b86-66f8cb72486d" (UID: "136dd621-a5c5-481a-9b86-66f8cb72486d"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.419607 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e945937-1e9d-465b-a68f-c755119ab6f9-kube-api-access-svj9w" (OuterVolumeSpecName: "kube-api-access-svj9w") pod "6e945937-1e9d-465b-a68f-c755119ab6f9" (UID: "6e945937-1e9d-465b-a68f-c755119ab6f9"). InnerVolumeSpecName "kube-api-access-svj9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.420048 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb362d0e-68b4-4d48-885c-a6894ac71e6e-kube-api-access-npm75" (OuterVolumeSpecName: "kube-api-access-npm75") pod "bb362d0e-68b4-4d48-885c-a6894ac71e6e" (UID: "bb362d0e-68b4-4d48-885c-a6894ac71e6e"). InnerVolumeSpecName "kube-api-access-npm75". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.420422 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/549beb55-bab4-4b1a-bf98-48de2e12956f-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "549beb55-bab4-4b1a-bf98-48de2e12956f" (UID: "549beb55-bab4-4b1a-bf98-48de2e12956f"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.420445 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "6e945937-1e9d-465b-a68f-c755119ab6f9" (UID: "6e945937-1e9d-465b-a68f-c755119ab6f9"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.420510 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da816a64-160f-4a5c-b945-b8fa467208d6-kube-api-access-4wc2m" (OuterVolumeSpecName: "kube-api-access-4wc2m") pod "da816a64-160f-4a5c-b945-b8fa467208d6" (UID: "da816a64-160f-4a5c-b945-b8fa467208d6"). InnerVolumeSpecName "kube-api-access-4wc2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.420542 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/136dd621-a5c5-481a-9b86-66f8cb72486d-kube-api-access-vp59z" (OuterVolumeSpecName: "kube-api-access-vp59z") pod "136dd621-a5c5-481a-9b86-66f8cb72486d" (UID: "136dd621-a5c5-481a-9b86-66f8cb72486d"). InnerVolumeSpecName "kube-api-access-vp59z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.422311 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4" (UID: "8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.422753 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-scripts" (OuterVolumeSpecName: "scripts") pod "6e945937-1e9d-465b-a68f-c755119ab6f9" (UID: "6e945937-1e9d-465b-a68f-c755119ab6f9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.423566 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-kube-api-access-jrsxt" (OuterVolumeSpecName: "kube-api-access-jrsxt") pod "8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4" (UID: "8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4"). InnerVolumeSpecName "kube-api-access-jrsxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.426712 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/549beb55-bab4-4b1a-bf98-48de2e12956f-kube-api-access-gxqmm" (OuterVolumeSpecName: "kube-api-access-gxqmm") pod "549beb55-bab4-4b1a-bf98-48de2e12956f" (UID: "549beb55-bab4-4b1a-bf98-48de2e12956f"). InnerVolumeSpecName "kube-api-access-gxqmm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.446992 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb362d0e-68b4-4d48-885c-a6894ac71e6e-config" (OuterVolumeSpecName: "config") pod "bb362d0e-68b4-4d48-885c-a6894ac71e6e" (UID: "bb362d0e-68b4-4d48-885c-a6894ac71e6e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.458817 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6e945937-1e9d-465b-a68f-c755119ab6f9" (UID: "6e945937-1e9d-465b-a68f-c755119ab6f9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.459342 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb362d0e-68b4-4d48-885c-a6894ac71e6e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bb362d0e-68b4-4d48-885c-a6894ac71e6e" (UID: "bb362d0e-68b4-4d48-885c-a6894ac71e6e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.472920 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-config" (OuterVolumeSpecName: "config") pod "da816a64-160f-4a5c-b945-b8fa467208d6" (UID: "da816a64-160f-4a5c-b945-b8fa467208d6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.483283 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "da816a64-160f-4a5c-b945-b8fa467208d6" (UID: "da816a64-160f-4a5c-b945-b8fa467208d6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.485653 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "da816a64-160f-4a5c-b945-b8fa467208d6" (UID: "da816a64-160f-4a5c-b945-b8fa467208d6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.485646 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-config-data" (OuterVolumeSpecName: "config-data") pod "6e945937-1e9d-465b-a68f-c755119ab6f9" (UID: "6e945937-1e9d-465b-a68f-c755119ab6f9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.486779 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "da816a64-160f-4a5c-b945-b8fa467208d6" (UID: "da816a64-160f-4a5c-b945-b8fa467208d6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.511895 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/136dd621-a5c5-481a-9b86-66f8cb72486d-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.511926 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrsxt\" (UniqueName: \"kubernetes.io/projected/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-kube-api-access-jrsxt\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.511939 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxqmm\" (UniqueName: \"kubernetes.io/projected/549beb55-bab4-4b1a-bf98-48de2e12956f-kube-api-access-gxqmm\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.511948 4948 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6e945937-1e9d-465b-a68f-c755119ab6f9-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.511956 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.511965 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.511984 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.511993 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/bb362d0e-68b4-4d48-885c-a6894ac71e6e-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512001 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e945937-1e9d-465b-a68f-c755119ab6f9-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512010 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/549beb55-bab4-4b1a-bf98-48de2e12956f-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512019 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/136dd621-a5c5-481a-9b86-66f8cb72486d-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512026 4948 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/549beb55-bab4-4b1a-bf98-48de2e12956f-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512034 4948 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/136dd621-a5c5-481a-9b86-66f8cb72486d-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512066 4948 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512076 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512084 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vp59z\" (UniqueName: \"kubernetes.io/projected/136dd621-a5c5-481a-9b86-66f8cb72486d-kube-api-access-vp59z\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512091 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/549beb55-bab4-4b1a-bf98-48de2e12956f-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512099 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svj9w\" (UniqueName: \"kubernetes.io/projected/6e945937-1e9d-465b-a68f-c755119ab6f9-kube-api-access-svj9w\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512107 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wc2m\" (UniqueName: \"kubernetes.io/projected/da816a64-160f-4a5c-b945-b8fa467208d6-kube-api-access-4wc2m\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512115 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512122 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512129 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512139 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/136dd621-a5c5-481a-9b86-66f8cb72486d-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512148 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512156 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/549beb55-bab4-4b1a-bf98-48de2e12956f-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512164 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512172 4948 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512180 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npm75\" (UniqueName: \"kubernetes.io/projected/bb362d0e-68b4-4d48-885c-a6894ac71e6e-kube-api-access-npm75\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512188 4948 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512197 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb362d0e-68b4-4d48-885c-a6894ac71e6e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.532293 4948 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.614374 4948 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.125909 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-g5zxh" event={"ID":"bb362d0e-68b4-4d48-885c-a6894ac71e6e","Type":"ContainerDied","Data":"fb0ac2219219fe66e5a42c9ab8b34f9c1f674fc235bce862e34dedbfa83dd480"} Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.125937 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-g5zxh" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.125953 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb0ac2219219fe66e5a42c9ab8b34f9c1f674fc235bce862e34dedbfa83dd480" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.127790 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68c9c4f5fc-2vtb4" event={"ID":"136dd621-a5c5-481a-9b86-66f8cb72486d","Type":"ContainerDied","Data":"45d6073bbc41d12c6fd4a83a0736dfc692eb915e2f643d37fcfcdd4574770ee8"} Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.128272 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.133310 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d578c747f-8f6hc" event={"ID":"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4","Type":"ContainerDied","Data":"3f991d1b028cfdcb1512441c7603f82fee2c2bba86ea276855760a70987a1d8f"} Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.133357 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.138509 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.138605 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-54d5f75fb9-s2659" event={"ID":"549beb55-bab4-4b1a-bf98-48de2e12956f","Type":"ContainerDied","Data":"27986056f94b762ad4c0ec91cfd1562122b907c3b47800656945a8588571afe8"} Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.138669 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.143086 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.222853 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-68c9c4f5fc-2vtb4"] Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.232860 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-68c9c4f5fc-2vtb4"] Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.251942 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-tkqzb"] Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.264702 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-tkqzb"] Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.286327 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-54d5f75fb9-s2659"] Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.295382 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-54d5f75fb9-s2659"] Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.313775 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7d578c747f-8f6hc"] Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.323245 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7d578c747f-8f6hc"] Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.332958 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.342088 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.352176 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 08:22:41 crc kubenswrapper[4948]: E0220 08:22:41.352611 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da816a64-160f-4a5c-b945-b8fa467208d6" containerName="dnsmasq-dns" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.352632 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="da816a64-160f-4a5c-b945-b8fa467208d6" containerName="dnsmasq-dns" Feb 20 08:22:41 crc kubenswrapper[4948]: E0220 08:22:41.352659 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e945937-1e9d-465b-a68f-c755119ab6f9" containerName="glance-httpd" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.352667 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e945937-1e9d-465b-a68f-c755119ab6f9" containerName="glance-httpd" Feb 20 08:22:41 crc kubenswrapper[4948]: E0220 08:22:41.352689 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb362d0e-68b4-4d48-885c-a6894ac71e6e" containerName="neutron-db-sync" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.352698 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb362d0e-68b4-4d48-885c-a6894ac71e6e" containerName="neutron-db-sync" Feb 20 08:22:41 crc kubenswrapper[4948]: E0220 08:22:41.352716 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da816a64-160f-4a5c-b945-b8fa467208d6" containerName="init" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.352723 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="da816a64-160f-4a5c-b945-b8fa467208d6" containerName="init" Feb 20 08:22:41 crc kubenswrapper[4948]: E0220 08:22:41.352738 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e945937-1e9d-465b-a68f-c755119ab6f9" containerName="glance-log" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.352745 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e945937-1e9d-465b-a68f-c755119ab6f9" containerName="glance-log" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.352951 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e945937-1e9d-465b-a68f-c755119ab6f9" containerName="glance-log" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.352990 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e945937-1e9d-465b-a68f-c755119ab6f9" containerName="glance-httpd" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.353001 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb362d0e-68b4-4d48-885c-a6894ac71e6e" containerName="neutron-db-sync" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.353010 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="da816a64-160f-4a5c-b945-b8fa467208d6" containerName="dnsmasq-dns" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.354129 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.356114 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.356513 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.359628 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.472043 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-v8n5p"] Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.473773 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.494440 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-v8n5p"] Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.527754 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6phw\" (UniqueName: \"kubernetes.io/projected/33b35605-1057-4e64-99a7-a7273351d6c2-kube-api-access-q6phw\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.527818 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.527876 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.527911 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/33b35605-1057-4e64-99a7-a7273351d6c2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.527939 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.527999 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.528028 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33b35605-1057-4e64-99a7-a7273351d6c2-logs\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.528079 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.581165 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-b5c5cc45d-7dcbq"] Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.582409 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.584808 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.585199 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.585573 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-4txrz" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.585949 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.603461 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b5c5cc45d-7dcbq"] Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.630657 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgbxd\" (UniqueName: \"kubernetes.io/projected/0cc4ba01-323f-4bcb-9bc3-d639514f6088-kube-api-access-sgbxd\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.630705 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.630738 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-dns-svc\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.630773 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6phw\" (UniqueName: \"kubernetes.io/projected/33b35605-1057-4e64-99a7-a7273351d6c2-kube-api-access-q6phw\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.630798 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.630854 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.630884 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/33b35605-1057-4e64-99a7-a7273351d6c2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.630902 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.630927 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-config\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.630961 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.631001 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33b35605-1057-4e64-99a7-a7273351d6c2-logs\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.631030 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.631064 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.631093 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.631722 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.632839 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/33b35605-1057-4e64-99a7-a7273351d6c2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.632879 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33b35605-1057-4e64-99a7-a7273351d6c2-logs\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.636120 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.638286 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.639498 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.649405 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.650954 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6phw\" (UniqueName: \"kubernetes.io/projected/33b35605-1057-4e64-99a7-a7273351d6c2-kube-api-access-q6phw\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.679025 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.733492 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="136dd621-a5c5-481a-9b86-66f8cb72486d" path="/var/lib/kubelet/pods/136dd621-a5c5-481a-9b86-66f8cb72486d/volumes" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.733929 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="549beb55-bab4-4b1a-bf98-48de2e12956f" path="/var/lib/kubelet/pods/549beb55-bab4-4b1a-bf98-48de2e12956f/volumes" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.734399 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e945937-1e9d-465b-a68f-c755119ab6f9" path="/var/lib/kubelet/pods/6e945937-1e9d-465b-a68f-c755119ab6f9/volumes" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.734881 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-ovndb-tls-certs\") pod \"neutron-b5c5cc45d-7dcbq\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.734952 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.735004 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khls2\" (UniqueName: \"kubernetes.io/projected/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-kube-api-access-khls2\") pod \"neutron-b5c5cc45d-7dcbq\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.735039 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.735099 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-config\") pod \"neutron-b5c5cc45d-7dcbq\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.735129 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-httpd-config\") pod \"neutron-b5c5cc45d-7dcbq\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.735153 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgbxd\" (UniqueName: \"kubernetes.io/projected/0cc4ba01-323f-4bcb-9bc3-d639514f6088-kube-api-access-sgbxd\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.735181 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.735228 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-dns-svc\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.735282 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-combined-ca-bundle\") pod \"neutron-b5c5cc45d-7dcbq\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.735317 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-config\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.736137 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4" path="/var/lib/kubelet/pods/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4/volumes" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.736548 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da816a64-160f-4a5c-b945-b8fa467208d6" path="/var/lib/kubelet/pods/da816a64-160f-4a5c-b945-b8fa467208d6/volumes" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.736569 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.736631 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.737212 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-config\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.737530 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-dns-svc\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.737678 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.754051 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgbxd\" (UniqueName: \"kubernetes.io/projected/0cc4ba01-323f-4bcb-9bc3-d639514f6088-kube-api-access-sgbxd\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.810650 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.837142 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khls2\" (UniqueName: \"kubernetes.io/projected/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-kube-api-access-khls2\") pod \"neutron-b5c5cc45d-7dcbq\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.837225 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-config\") pod \"neutron-b5c5cc45d-7dcbq\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.837262 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-httpd-config\") pod \"neutron-b5c5cc45d-7dcbq\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.837387 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-combined-ca-bundle\") pod \"neutron-b5c5cc45d-7dcbq\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.837439 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-ovndb-tls-certs\") pod \"neutron-b5c5cc45d-7dcbq\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.843164 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-ovndb-tls-certs\") pod \"neutron-b5c5cc45d-7dcbq\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.844152 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-httpd-config\") pod \"neutron-b5c5cc45d-7dcbq\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.847818 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-config\") pod \"neutron-b5c5cc45d-7dcbq\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.852685 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-combined-ca-bundle\") pod \"neutron-b5c5cc45d-7dcbq\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.859647 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khls2\" (UniqueName: \"kubernetes.io/projected/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-kube-api-access-khls2\") pod \"neutron-b5c5cc45d-7dcbq\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.908649 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.978604 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 20 08:22:42 crc kubenswrapper[4948]: E0220 08:22:42.059487 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Feb 20 08:22:42 crc kubenswrapper[4948]: E0220 08:22:42.059678 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fwtv5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-xt2zq_openstack(f706b741-86f9-49a3-95d5-85ee62eb3668): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 08:22:42 crc kubenswrapper[4948]: E0220 08:22:42.060815 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-xt2zq" podUID="f706b741-86f9-49a3-95d5-85ee62eb3668" Feb 20 08:22:42 crc kubenswrapper[4948]: E0220 08:22:42.177132 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-xt2zq" podUID="f706b741-86f9-49a3-95d5-85ee62eb3668" Feb 20 08:22:42 crc kubenswrapper[4948]: W0220 08:22:42.675476 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b9a072a_a040_414c_96c7_cf1454148745.slice/crio-6dcfb76c7284437f090e639d57144c71edd62b4bab7242fd64ad46e096ae5a7b WatchSource:0}: Error finding container 6dcfb76c7284437f090e639d57144c71edd62b4bab7242fd64ad46e096ae5a7b: Status 404 returned error can't find the container with id 6dcfb76c7284437f090e639d57144c71edd62b4bab7242fd64ad46e096ae5a7b Feb 20 08:22:42 crc kubenswrapper[4948]: I0220 08:22:42.680594 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-685c4db87b-rw57h"] Feb 20 08:22:42 crc kubenswrapper[4948]: I0220 08:22:42.772116 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g27hk"] Feb 20 08:22:42 crc kubenswrapper[4948]: I0220 08:22:42.797821 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-66b4d4cb8-hgmkd"] Feb 20 08:22:42 crc kubenswrapper[4948]: I0220 08:22:42.842062 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:22:42 crc kubenswrapper[4948]: W0220 08:22:42.852227 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0f82f51_8a20_402d_8861_6c57a4e5af0d.slice/crio-70cd0cac387b5f659d055b8820f12fcea594407eebec35ab497b573fb1f4db5a WatchSource:0}: Error finding container 70cd0cac387b5f659d055b8820f12fcea594407eebec35ab497b573fb1f4db5a: Status 404 returned error can't find the container with id 70cd0cac387b5f659d055b8820f12fcea594407eebec35ab497b573fb1f4db5a Feb 20 08:22:42 crc kubenswrapper[4948]: I0220 08:22:42.883468 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hh5s2"] Feb 20 08:22:42 crc kubenswrapper[4948]: W0220 08:22:42.898317 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd84a71b6_7503_4150_a5c4_4579b08a669a.slice/crio-21944c755e2a58a194f7e12be68919b81d0c021ee6380fb185cda3cec0fcee37 WatchSource:0}: Error finding container 21944c755e2a58a194f7e12be68919b81d0c021ee6380fb185cda3cec0fcee37: Status 404 returned error can't find the container with id 21944c755e2a58a194f7e12be68919b81d0c021ee6380fb185cda3cec0fcee37 Feb 20 08:22:42 crc kubenswrapper[4948]: I0220 08:22:42.903703 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-v29fx"] Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.029088 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-v8n5p"] Feb 20 08:22:43 crc kubenswrapper[4948]: W0220 08:22:43.048032 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0cc4ba01_323f_4bcb_9bc3_d639514f6088.slice/crio-5ba21a5e281f6be663a46ae25fc6e9402ee9bcd0e6caec5c0c3b4658222b76c1 WatchSource:0}: Error finding container 5ba21a5e281f6be663a46ae25fc6e9402ee9bcd0e6caec5c0c3b4658222b76c1: Status 404 returned error can't find the container with id 5ba21a5e281f6be663a46ae25fc6e9402ee9bcd0e6caec5c0c3b4658222b76c1 Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.187480 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6mz7x" event={"ID":"1227b752-8a3c-4c01-91bf-0662c1b1a231","Type":"ContainerStarted","Data":"b8d5cd74c5cbd29dae34c7b1cedbde475f451d2de997a59b511db6cc61ba0a29"} Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.203114 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66b4d4cb8-hgmkd" event={"ID":"d1046503-c6d3-49bf-9523-2f897fcb82ba","Type":"ContainerStarted","Data":"0fc9151c9d2a99b81f2c9346beebf0499347807bacbd7c95da55982c7422c38e"} Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.224417 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.227846 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-6mz7x" podStartSLOduration=5.800811674 podStartE2EDuration="29.2278275s" podCreationTimestamp="2026-02-20 08:22:14 +0000 UTC" firstStartedPulling="2026-02-20 08:22:16.65605074 +0000 UTC m=+985.630545560" lastFinishedPulling="2026-02-20 08:22:40.083066546 +0000 UTC m=+1009.057561386" observedRunningTime="2026-02-20 08:22:43.224868687 +0000 UTC m=+1012.199363507" watchObservedRunningTime="2026-02-20 08:22:43.2278275 +0000 UTC m=+1012.202322320" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.244168 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89b49c50-bae7-47fe-b85f-7a718953933d","Type":"ContainerStarted","Data":"f21ea8eddb7fd63986bb5357a809bc7c94ba2e30ea5976b9795bdbd7d103ba56"} Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.259945 4948 generic.go:334] "Generic (PLEG): container finished" podID="d6729921-0a12-46a0-824f-810df88ae426" containerID="c30e42a1011ef0a11a3bd3a8a1babce0958085ece91112de32b9f174e1d1c87a" exitCode=0 Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.260031 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g27hk" event={"ID":"d6729921-0a12-46a0-824f-810df88ae426","Type":"ContainerDied","Data":"c30e42a1011ef0a11a3bd3a8a1babce0958085ece91112de32b9f174e1d1c87a"} Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.260060 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g27hk" event={"ID":"d6729921-0a12-46a0-824f-810df88ae426","Type":"ContainerStarted","Data":"b7570c60dc7ea9b2dda7deeb7d6ef87e7350a05f972449d37dfa272d75ce2941"} Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.272664 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hh5s2" event={"ID":"426b37f9-4acb-4d28-a55c-3fcde2600002","Type":"ContainerStarted","Data":"6c8a403f4e4cb34b74e5da72f8bdf18b9a2a922052dce8500240a25dba684641"} Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.302253 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v29fx" event={"ID":"d84a71b6-7503-4150-a5c4-4579b08a669a","Type":"ContainerStarted","Data":"21944c755e2a58a194f7e12be68919b81d0c021ee6380fb185cda3cec0fcee37"} Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.307192 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e0f82f51-8a20-402d-8861-6c57a4e5af0d","Type":"ContainerStarted","Data":"70cd0cac387b5f659d055b8820f12fcea594407eebec35ab497b573fb1f4db5a"} Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.314320 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" event={"ID":"0cc4ba01-323f-4bcb-9bc3-d639514f6088","Type":"ContainerStarted","Data":"5ba21a5e281f6be663a46ae25fc6e9402ee9bcd0e6caec5c0c3b4658222b76c1"} Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.320863 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-685c4db87b-rw57h" event={"ID":"7b9a072a-a040-414c-96c7-cf1454148745","Type":"ContainerStarted","Data":"6dcfb76c7284437f090e639d57144c71edd62b4bab7242fd64ad46e096ae5a7b"} Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.548894 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-685bb7f5c-jd9nm"] Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.556874 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.564932 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.566202 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.575810 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-685bb7f5c-jd9nm"] Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.705591 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fh687\" (UniqueName: \"kubernetes.io/projected/e0661780-ca2b-42ae-be25-5d0e914fbb49-kube-api-access-fh687\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.706202 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-combined-ca-bundle\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.706360 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-public-tls-certs\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.706382 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-ovndb-tls-certs\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.706909 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-config\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.706950 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-internal-tls-certs\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.707630 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-httpd-config\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.808739 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-combined-ca-bundle\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.808812 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-public-tls-certs\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.808832 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-ovndb-tls-certs\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.810208 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-config\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.810328 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-internal-tls-certs\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.810608 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-httpd-config\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.810633 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fh687\" (UniqueName: \"kubernetes.io/projected/e0661780-ca2b-42ae-be25-5d0e914fbb49-kube-api-access-fh687\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.812646 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-public-tls-certs\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.812912 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-combined-ca-bundle\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.817554 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-ovndb-tls-certs\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.818235 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-config\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.825815 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-internal-tls-certs\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.826512 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-httpd-config\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.830687 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fh687\" (UniqueName: \"kubernetes.io/projected/e0661780-ca2b-42ae-be25-5d0e914fbb49-kube-api-access-fh687\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.923624 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:44 crc kubenswrapper[4948]: I0220 08:22:44.076452 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b5c5cc45d-7dcbq"] Feb 20 08:22:44 crc kubenswrapper[4948]: W0220 08:22:44.114685 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd1b85e2_1b5f_4993_bc66_72daaf6aa79d.slice/crio-4c4522a0ae94ffaed4b0ddc1a755c0498d0a11f95af5418408b70c0351708567 WatchSource:0}: Error finding container 4c4522a0ae94ffaed4b0ddc1a755c0498d0a11f95af5418408b70c0351708567: Status 404 returned error can't find the container with id 4c4522a0ae94ffaed4b0ddc1a755c0498d0a11f95af5418408b70c0351708567 Feb 20 08:22:44 crc kubenswrapper[4948]: I0220 08:22:44.344115 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-685c4db87b-rw57h" event={"ID":"7b9a072a-a040-414c-96c7-cf1454148745","Type":"ContainerStarted","Data":"827809de35c30bbe5e9607a26506e33891d08d97441e79c25a85c226aa503507"} Feb 20 08:22:44 crc kubenswrapper[4948]: I0220 08:22:44.345444 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-685c4db87b-rw57h" event={"ID":"7b9a072a-a040-414c-96c7-cf1454148745","Type":"ContainerStarted","Data":"f855ae3e2aaeae487e0cde870ef0f634ece3468bf55857211d1d8d871cc6a217"} Feb 20 08:22:44 crc kubenswrapper[4948]: I0220 08:22:44.348308 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e0f82f51-8a20-402d-8861-6c57a4e5af0d","Type":"ContainerStarted","Data":"7020629c426e657a58b5859f74b5d55f5c4fb964822922a0eb88d95415abe7d8"} Feb 20 08:22:44 crc kubenswrapper[4948]: I0220 08:22:44.354251 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66b4d4cb8-hgmkd" event={"ID":"d1046503-c6d3-49bf-9523-2f897fcb82ba","Type":"ContainerStarted","Data":"577865ca69f0d8457d617948e5cac2be0a124e4620edc60df1395130de8a2334"} Feb 20 08:22:44 crc kubenswrapper[4948]: I0220 08:22:44.357391 4948 generic.go:334] "Generic (PLEG): container finished" podID="0cc4ba01-323f-4bcb-9bc3-d639514f6088" containerID="7121b5f51a9b7787780e4d13a76cf2b72d32878e2e19ab652f2bed18d9c68743" exitCode=0 Feb 20 08:22:44 crc kubenswrapper[4948]: I0220 08:22:44.357826 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" event={"ID":"0cc4ba01-323f-4bcb-9bc3-d639514f6088","Type":"ContainerDied","Data":"7121b5f51a9b7787780e4d13a76cf2b72d32878e2e19ab652f2bed18d9c68743"} Feb 20 08:22:44 crc kubenswrapper[4948]: I0220 08:22:44.364514 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b5c5cc45d-7dcbq" event={"ID":"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d","Type":"ContainerStarted","Data":"4c4522a0ae94ffaed4b0ddc1a755c0498d0a11f95af5418408b70c0351708567"} Feb 20 08:22:44 crc kubenswrapper[4948]: I0220 08:22:44.370455 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-685c4db87b-rw57h" podStartSLOduration=18.836466751 podStartE2EDuration="19.370432985s" podCreationTimestamp="2026-02-20 08:22:25 +0000 UTC" firstStartedPulling="2026-02-20 08:22:42.677224764 +0000 UTC m=+1011.651719584" lastFinishedPulling="2026-02-20 08:22:43.211190998 +0000 UTC m=+1012.185685818" observedRunningTime="2026-02-20 08:22:44.364765685 +0000 UTC m=+1013.339260505" watchObservedRunningTime="2026-02-20 08:22:44.370432985 +0000 UTC m=+1013.344927805" Feb 20 08:22:44 crc kubenswrapper[4948]: I0220 08:22:44.371291 4948 generic.go:334] "Generic (PLEG): container finished" podID="426b37f9-4acb-4d28-a55c-3fcde2600002" containerID="62e9964b1041b8a3ee083b7749dd1ff49cbef0514cc265f8b36923ced11f9863" exitCode=0 Feb 20 08:22:44 crc kubenswrapper[4948]: I0220 08:22:44.371373 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hh5s2" event={"ID":"426b37f9-4acb-4d28-a55c-3fcde2600002","Type":"ContainerDied","Data":"62e9964b1041b8a3ee083b7749dd1ff49cbef0514cc265f8b36923ced11f9863"} Feb 20 08:22:44 crc kubenswrapper[4948]: I0220 08:22:44.373792 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v29fx" event={"ID":"d84a71b6-7503-4150-a5c4-4579b08a669a","Type":"ContainerStarted","Data":"87c2cdb1a2b7f5c38a01dd7f884ce5938563a33612ebaf51f20ceac5532f97f2"} Feb 20 08:22:44 crc kubenswrapper[4948]: I0220 08:22:44.394335 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"33b35605-1057-4e64-99a7-a7273351d6c2","Type":"ContainerStarted","Data":"9232d9b3b9f362e1b01dbc32c5300081164bdbed5e40b2c4cd59a5e4f4b28166"} Feb 20 08:22:44 crc kubenswrapper[4948]: I0220 08:22:44.394385 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"33b35605-1057-4e64-99a7-a7273351d6c2","Type":"ContainerStarted","Data":"3abd35d213010021a69fde90dd757e9181aec2b84f4e7f71376ea0edef4300a5"} Feb 20 08:22:44 crc kubenswrapper[4948]: I0220 08:22:44.412173 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-v29fx" podStartSLOduration=16.412158248 podStartE2EDuration="16.412158248s" podCreationTimestamp="2026-02-20 08:22:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:22:44.408938268 +0000 UTC m=+1013.383433108" watchObservedRunningTime="2026-02-20 08:22:44.412158248 +0000 UTC m=+1013.386653068" Feb 20 08:22:44 crc kubenswrapper[4948]: I0220 08:22:44.631493 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-685bb7f5c-jd9nm"] Feb 20 08:22:44 crc kubenswrapper[4948]: I0220 08:22:44.851277 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-tkqzb" podUID="da816a64-160f-4a5c-b945-b8fa467208d6" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: i/o timeout" Feb 20 08:22:45 crc kubenswrapper[4948]: I0220 08:22:45.404777 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-685bb7f5c-jd9nm" event={"ID":"e0661780-ca2b-42ae-be25-5d0e914fbb49","Type":"ContainerStarted","Data":"4885fc2e4d8fe9fc0c29512f6a5fd4f1b7a87efbf9c10fe17b8bf4fe7f1ca02a"} Feb 20 08:22:45 crc kubenswrapper[4948]: I0220 08:22:45.407096 4948 generic.go:334] "Generic (PLEG): container finished" podID="1227b752-8a3c-4c01-91bf-0662c1b1a231" containerID="b8d5cd74c5cbd29dae34c7b1cedbde475f451d2de997a59b511db6cc61ba0a29" exitCode=0 Feb 20 08:22:45 crc kubenswrapper[4948]: I0220 08:22:45.407156 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6mz7x" event={"ID":"1227b752-8a3c-4c01-91bf-0662c1b1a231","Type":"ContainerDied","Data":"b8d5cd74c5cbd29dae34c7b1cedbde475f451d2de997a59b511db6cc61ba0a29"} Feb 20 08:22:45 crc kubenswrapper[4948]: I0220 08:22:45.409489 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66b4d4cb8-hgmkd" event={"ID":"d1046503-c6d3-49bf-9523-2f897fcb82ba","Type":"ContainerStarted","Data":"a1b60a312d88b84db5225c1aff5290457499b2775b6fffec7ca3246896b28523"} Feb 20 08:22:45 crc kubenswrapper[4948]: I0220 08:22:45.413413 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" event={"ID":"0cc4ba01-323f-4bcb-9bc3-d639514f6088","Type":"ContainerStarted","Data":"ce123808c7ffa9e9792ff9cd3fa77559a772ebe4ee4f08ad46bdfe3f21ea2386"} Feb 20 08:22:45 crc kubenswrapper[4948]: I0220 08:22:45.413575 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:45 crc kubenswrapper[4948]: I0220 08:22:45.415048 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b5c5cc45d-7dcbq" event={"ID":"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d","Type":"ContainerStarted","Data":"fceb35ea8befbc40c2edb25ea5ef78a9c4a759e83b13e451735a1e2ccf6a99a8"} Feb 20 08:22:45 crc kubenswrapper[4948]: I0220 08:22:45.418263 4948 generic.go:334] "Generic (PLEG): container finished" podID="426b37f9-4acb-4d28-a55c-3fcde2600002" containerID="2d58a1beb2801a36e08660bfb99639b7abb52f4d18887bdac3f68efb75c8316d" exitCode=0 Feb 20 08:22:45 crc kubenswrapper[4948]: I0220 08:22:45.418446 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hh5s2" event={"ID":"426b37f9-4acb-4d28-a55c-3fcde2600002","Type":"ContainerDied","Data":"2d58a1beb2801a36e08660bfb99639b7abb52f4d18887bdac3f68efb75c8316d"} Feb 20 08:22:45 crc kubenswrapper[4948]: I0220 08:22:45.450362 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" podStartSLOduration=4.45034643 podStartE2EDuration="4.45034643s" podCreationTimestamp="2026-02-20 08:22:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:22:45.4475054 +0000 UTC m=+1014.422000220" watchObservedRunningTime="2026-02-20 08:22:45.45034643 +0000 UTC m=+1014.424841250" Feb 20 08:22:45 crc kubenswrapper[4948]: I0220 08:22:45.477270 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:45 crc kubenswrapper[4948]: I0220 08:22:45.477314 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:45 crc kubenswrapper[4948]: I0220 08:22:45.498401 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-66b4d4cb8-hgmkd" podStartSLOduration=19.956081589 podStartE2EDuration="20.498382519s" podCreationTimestamp="2026-02-20 08:22:25 +0000 UTC" firstStartedPulling="2026-02-20 08:22:42.822392586 +0000 UTC m=+1011.796887406" lastFinishedPulling="2026-02-20 08:22:43.364693516 +0000 UTC m=+1012.339188336" observedRunningTime="2026-02-20 08:22:45.492868092 +0000 UTC m=+1014.467362922" watchObservedRunningTime="2026-02-20 08:22:45.498382519 +0000 UTC m=+1014.472877339" Feb 20 08:22:45 crc kubenswrapper[4948]: I0220 08:22:45.643332 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:45 crc kubenswrapper[4948]: I0220 08:22:45.643510 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:46 crc kubenswrapper[4948]: I0220 08:22:46.874570 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:46 crc kubenswrapper[4948]: I0220 08:22:46.976446 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-config-data\") pod \"1227b752-8a3c-4c01-91bf-0662c1b1a231\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " Feb 20 08:22:46 crc kubenswrapper[4948]: I0220 08:22:46.976536 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1227b752-8a3c-4c01-91bf-0662c1b1a231-logs\") pod \"1227b752-8a3c-4c01-91bf-0662c1b1a231\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " Feb 20 08:22:46 crc kubenswrapper[4948]: I0220 08:22:46.976630 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rtp8\" (UniqueName: \"kubernetes.io/projected/1227b752-8a3c-4c01-91bf-0662c1b1a231-kube-api-access-5rtp8\") pod \"1227b752-8a3c-4c01-91bf-0662c1b1a231\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " Feb 20 08:22:46 crc kubenswrapper[4948]: I0220 08:22:46.976683 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-scripts\") pod \"1227b752-8a3c-4c01-91bf-0662c1b1a231\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " Feb 20 08:22:46 crc kubenswrapper[4948]: I0220 08:22:46.976805 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-combined-ca-bundle\") pod \"1227b752-8a3c-4c01-91bf-0662c1b1a231\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " Feb 20 08:22:46 crc kubenswrapper[4948]: I0220 08:22:46.976933 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1227b752-8a3c-4c01-91bf-0662c1b1a231-logs" (OuterVolumeSpecName: "logs") pod "1227b752-8a3c-4c01-91bf-0662c1b1a231" (UID: "1227b752-8a3c-4c01-91bf-0662c1b1a231"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:22:46 crc kubenswrapper[4948]: I0220 08:22:46.977273 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1227b752-8a3c-4c01-91bf-0662c1b1a231-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:46 crc kubenswrapper[4948]: I0220 08:22:46.987392 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1227b752-8a3c-4c01-91bf-0662c1b1a231-kube-api-access-5rtp8" (OuterVolumeSpecName: "kube-api-access-5rtp8") pod "1227b752-8a3c-4c01-91bf-0662c1b1a231" (UID: "1227b752-8a3c-4c01-91bf-0662c1b1a231"). InnerVolumeSpecName "kube-api-access-5rtp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:46 crc kubenswrapper[4948]: I0220 08:22:46.992302 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-scripts" (OuterVolumeSpecName: "scripts") pod "1227b752-8a3c-4c01-91bf-0662c1b1a231" (UID: "1227b752-8a3c-4c01-91bf-0662c1b1a231"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.018304 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-config-data" (OuterVolumeSpecName: "config-data") pod "1227b752-8a3c-4c01-91bf-0662c1b1a231" (UID: "1227b752-8a3c-4c01-91bf-0662c1b1a231"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.039744 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1227b752-8a3c-4c01-91bf-0662c1b1a231" (UID: "1227b752-8a3c-4c01-91bf-0662c1b1a231"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.078448 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.078481 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.078495 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rtp8\" (UniqueName: \"kubernetes.io/projected/1227b752-8a3c-4c01-91bf-0662c1b1a231-kube-api-access-5rtp8\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.078507 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.484665 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6mz7x" event={"ID":"1227b752-8a3c-4c01-91bf-0662c1b1a231","Type":"ContainerDied","Data":"5b0f5c257994604a44a9cf0e2c06abc610dcd1eb5c3d34cfd415bfd74671d08f"} Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.484708 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b0f5c257994604a44a9cf0e2c06abc610dcd1eb5c3d34cfd415bfd74671d08f" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.484793 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.487479 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89b49c50-bae7-47fe-b85f-7a718953933d","Type":"ContainerStarted","Data":"68df1bb1b13a5e62c745d68de5298b184c0cdcaa037a022eedeb800b70856493"} Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.496895 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g27hk" event={"ID":"d6729921-0a12-46a0-824f-810df88ae426","Type":"ContainerStarted","Data":"bdb7ba6b04df20596c79e2a31fb3e44aca83f1668de6adef1942e6b1ba6fa793"} Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.563451 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-d46988dcb-rn6qs"] Feb 20 08:22:47 crc kubenswrapper[4948]: E0220 08:22:47.563850 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1227b752-8a3c-4c01-91bf-0662c1b1a231" containerName="placement-db-sync" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.563868 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="1227b752-8a3c-4c01-91bf-0662c1b1a231" containerName="placement-db-sync" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.564073 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="1227b752-8a3c-4c01-91bf-0662c1b1a231" containerName="placement-db-sync" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.572666 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.584149 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.584301 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.584585 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.584838 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-2fq52" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.585220 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.588494 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-d46988dcb-rn6qs"] Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.696404 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-internal-tls-certs\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.696904 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bpmf\" (UniqueName: \"kubernetes.io/projected/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-kube-api-access-2bpmf\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.696951 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-public-tls-certs\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.697021 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-combined-ca-bundle\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.697892 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-config-data\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.698009 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-scripts\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.698227 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-logs\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.799570 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-config-data\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.799621 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-scripts\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.799642 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-logs\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.799685 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-internal-tls-certs\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.799740 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bpmf\" (UniqueName: \"kubernetes.io/projected/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-kube-api-access-2bpmf\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.799763 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-public-tls-certs\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.799778 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-combined-ca-bundle\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.800184 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-logs\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.811381 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-internal-tls-certs\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.815595 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-public-tls-certs\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.815739 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-combined-ca-bundle\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.815893 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-config-data\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.817271 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-scripts\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.820798 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bpmf\" (UniqueName: \"kubernetes.io/projected/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-kube-api-access-2bpmf\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.891354 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:48 crc kubenswrapper[4948]: I0220 08:22:48.507029 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"33b35605-1057-4e64-99a7-a7273351d6c2","Type":"ContainerStarted","Data":"08ec8e2a4374c5a50916bcc6e4fbc0d7957aac63564273dd75ed243c1b2a0679"} Feb 20 08:22:48 crc kubenswrapper[4948]: I0220 08:22:48.511199 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e0f82f51-8a20-402d-8861-6c57a4e5af0d","Type":"ContainerStarted","Data":"1a5b270055c8b2b5a29d54a228d2fd2e9bdac72aaebc130807e4057de5afdb1a"} Feb 20 08:22:48 crc kubenswrapper[4948]: I0220 08:22:48.511348 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e0f82f51-8a20-402d-8861-6c57a4e5af0d" containerName="glance-log" containerID="cri-o://7020629c426e657a58b5859f74b5d55f5c4fb964822922a0eb88d95415abe7d8" gracePeriod=30 Feb 20 08:22:48 crc kubenswrapper[4948]: I0220 08:22:48.511409 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e0f82f51-8a20-402d-8861-6c57a4e5af0d" containerName="glance-httpd" containerID="cri-o://1a5b270055c8b2b5a29d54a228d2fd2e9bdac72aaebc130807e4057de5afdb1a" gracePeriod=30 Feb 20 08:22:48 crc kubenswrapper[4948]: I0220 08:22:48.520894 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b5c5cc45d-7dcbq" event={"ID":"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d","Type":"ContainerStarted","Data":"4fc7cff150a17fca1672ea7d9eae41c85a175b0d1f41224639307a4d7025cd12"} Feb 20 08:22:48 crc kubenswrapper[4948]: I0220 08:22:48.521116 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:48 crc kubenswrapper[4948]: I0220 08:22:48.524605 4948 generic.go:334] "Generic (PLEG): container finished" podID="d6729921-0a12-46a0-824f-810df88ae426" containerID="bdb7ba6b04df20596c79e2a31fb3e44aca83f1668de6adef1942e6b1ba6fa793" exitCode=0 Feb 20 08:22:48 crc kubenswrapper[4948]: I0220 08:22:48.524728 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g27hk" event={"ID":"d6729921-0a12-46a0-824f-810df88ae426","Type":"ContainerDied","Data":"bdb7ba6b04df20596c79e2a31fb3e44aca83f1668de6adef1942e6b1ba6fa793"} Feb 20 08:22:48 crc kubenswrapper[4948]: I0220 08:22:48.525924 4948 generic.go:334] "Generic (PLEG): container finished" podID="d84a71b6-7503-4150-a5c4-4579b08a669a" containerID="87c2cdb1a2b7f5c38a01dd7f884ce5938563a33612ebaf51f20ceac5532f97f2" exitCode=0 Feb 20 08:22:48 crc kubenswrapper[4948]: I0220 08:22:48.526055 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v29fx" event={"ID":"d84a71b6-7503-4150-a5c4-4579b08a669a","Type":"ContainerDied","Data":"87c2cdb1a2b7f5c38a01dd7f884ce5938563a33612ebaf51f20ceac5532f97f2"} Feb 20 08:22:48 crc kubenswrapper[4948]: I0220 08:22:48.534953 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.534935674 podStartE2EDuration="7.534935674s" podCreationTimestamp="2026-02-20 08:22:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:22:48.530491234 +0000 UTC m=+1017.504986054" watchObservedRunningTime="2026-02-20 08:22:48.534935674 +0000 UTC m=+1017.509430494" Feb 20 08:22:48 crc kubenswrapper[4948]: I0220 08:22:48.536594 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-685bb7f5c-jd9nm" event={"ID":"e0661780-ca2b-42ae-be25-5d0e914fbb49","Type":"ContainerStarted","Data":"1ad25f5bfef4b0fa06a939cc9fb8e5bfa3dad488d3265a3cd5fa655c097dda5b"} Feb 20 08:22:48 crc kubenswrapper[4948]: I0220 08:22:48.583011 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=26.582960403 podStartE2EDuration="26.582960403s" podCreationTimestamp="2026-02-20 08:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:22:48.571639442 +0000 UTC m=+1017.546134262" watchObservedRunningTime="2026-02-20 08:22:48.582960403 +0000 UTC m=+1017.557455223" Feb 20 08:22:48 crc kubenswrapper[4948]: I0220 08:22:48.612133 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-b5c5cc45d-7dcbq" podStartSLOduration=7.612108314 podStartE2EDuration="7.612108314s" podCreationTimestamp="2026-02-20 08:22:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:22:48.599041891 +0000 UTC m=+1017.573536711" watchObservedRunningTime="2026-02-20 08:22:48.612108314 +0000 UTC m=+1017.586603144" Feb 20 08:22:49 crc kubenswrapper[4948]: I0220 08:22:49.335448 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-d46988dcb-rn6qs"] Feb 20 08:22:49 crc kubenswrapper[4948]: W0220 08:22:49.344491 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8b1ee525_1a6e_461d_aa4d_178ab601ec9d.slice/crio-101e3a4bd62000e7fe0c14babf5f059dc5dad0bc38ccadaa7bd0301c2ad9ec7f WatchSource:0}: Error finding container 101e3a4bd62000e7fe0c14babf5f059dc5dad0bc38ccadaa7bd0301c2ad9ec7f: Status 404 returned error can't find the container with id 101e3a4bd62000e7fe0c14babf5f059dc5dad0bc38ccadaa7bd0301c2ad9ec7f Feb 20 08:22:49 crc kubenswrapper[4948]: I0220 08:22:49.552588 4948 generic.go:334] "Generic (PLEG): container finished" podID="e0f82f51-8a20-402d-8861-6c57a4e5af0d" containerID="1a5b270055c8b2b5a29d54a228d2fd2e9bdac72aaebc130807e4057de5afdb1a" exitCode=0 Feb 20 08:22:49 crc kubenswrapper[4948]: I0220 08:22:49.553021 4948 generic.go:334] "Generic (PLEG): container finished" podID="e0f82f51-8a20-402d-8861-6c57a4e5af0d" containerID="7020629c426e657a58b5859f74b5d55f5c4fb964822922a0eb88d95415abe7d8" exitCode=143 Feb 20 08:22:49 crc kubenswrapper[4948]: I0220 08:22:49.553078 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e0f82f51-8a20-402d-8861-6c57a4e5af0d","Type":"ContainerDied","Data":"1a5b270055c8b2b5a29d54a228d2fd2e9bdac72aaebc130807e4057de5afdb1a"} Feb 20 08:22:49 crc kubenswrapper[4948]: I0220 08:22:49.553111 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e0f82f51-8a20-402d-8861-6c57a4e5af0d","Type":"ContainerDied","Data":"7020629c426e657a58b5859f74b5d55f5c4fb964822922a0eb88d95415abe7d8"} Feb 20 08:22:49 crc kubenswrapper[4948]: I0220 08:22:49.556805 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d46988dcb-rn6qs" event={"ID":"8b1ee525-1a6e-461d-aa4d-178ab601ec9d","Type":"ContainerStarted","Data":"101e3a4bd62000e7fe0c14babf5f059dc5dad0bc38ccadaa7bd0301c2ad9ec7f"} Feb 20 08:22:49 crc kubenswrapper[4948]: I0220 08:22:49.562191 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-685bb7f5c-jd9nm" event={"ID":"e0661780-ca2b-42ae-be25-5d0e914fbb49","Type":"ContainerStarted","Data":"d2ca1451e75be7daafc6907a8305baa229ed6441f1f79895bd2df00e5c17bfa7"} Feb 20 08:22:49 crc kubenswrapper[4948]: I0220 08:22:49.584445 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-685bb7f5c-jd9nm" podStartSLOduration=6.584426876 podStartE2EDuration="6.584426876s" podCreationTimestamp="2026-02-20 08:22:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:22:49.583555925 +0000 UTC m=+1018.558050735" watchObservedRunningTime="2026-02-20 08:22:49.584426876 +0000 UTC m=+1018.558921696" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.051082 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.060301 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.152083 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-credential-keys\") pod \"d84a71b6-7503-4150-a5c4-4579b08a669a\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.152146 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-config-data\") pod \"d84a71b6-7503-4150-a5c4-4579b08a669a\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.152416 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-scripts\") pod \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.152463 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5zcb\" (UniqueName: \"kubernetes.io/projected/e0f82f51-8a20-402d-8861-6c57a4e5af0d-kube-api-access-q5zcb\") pod \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.152500 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-scripts\") pod \"d84a71b6-7503-4150-a5c4-4579b08a669a\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.152522 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e0f82f51-8a20-402d-8861-6c57a4e5af0d-httpd-run\") pod \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.152546 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-combined-ca-bundle\") pod \"d84a71b6-7503-4150-a5c4-4579b08a669a\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.152566 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-config-data\") pod \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.152609 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-fernet-keys\") pod \"d84a71b6-7503-4150-a5c4-4579b08a669a\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.152630 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.152668 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-combined-ca-bundle\") pod \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.152701 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jr9d\" (UniqueName: \"kubernetes.io/projected/d84a71b6-7503-4150-a5c4-4579b08a669a-kube-api-access-6jr9d\") pod \"d84a71b6-7503-4150-a5c4-4579b08a669a\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.152718 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0f82f51-8a20-402d-8861-6c57a4e5af0d-logs\") pod \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.154009 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0f82f51-8a20-402d-8861-6c57a4e5af0d-logs" (OuterVolumeSpecName: "logs") pod "e0f82f51-8a20-402d-8861-6c57a4e5af0d" (UID: "e0f82f51-8a20-402d-8861-6c57a4e5af0d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.165818 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "d84a71b6-7503-4150-a5c4-4579b08a669a" (UID: "d84a71b6-7503-4150-a5c4-4579b08a669a"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.166253 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "e0f82f51-8a20-402d-8861-6c57a4e5af0d" (UID: "e0f82f51-8a20-402d-8861-6c57a4e5af0d"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.173147 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0f82f51-8a20-402d-8861-6c57a4e5af0d-kube-api-access-q5zcb" (OuterVolumeSpecName: "kube-api-access-q5zcb") pod "e0f82f51-8a20-402d-8861-6c57a4e5af0d" (UID: "e0f82f51-8a20-402d-8861-6c57a4e5af0d"). InnerVolumeSpecName "kube-api-access-q5zcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.178348 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0f82f51-8a20-402d-8861-6c57a4e5af0d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e0f82f51-8a20-402d-8861-6c57a4e5af0d" (UID: "e0f82f51-8a20-402d-8861-6c57a4e5af0d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.178347 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "d84a71b6-7503-4150-a5c4-4579b08a669a" (UID: "d84a71b6-7503-4150-a5c4-4579b08a669a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.180191 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d84a71b6-7503-4150-a5c4-4579b08a669a-kube-api-access-6jr9d" (OuterVolumeSpecName: "kube-api-access-6jr9d") pod "d84a71b6-7503-4150-a5c4-4579b08a669a" (UID: "d84a71b6-7503-4150-a5c4-4579b08a669a"). InnerVolumeSpecName "kube-api-access-6jr9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.202477 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-scripts" (OuterVolumeSpecName: "scripts") pod "d84a71b6-7503-4150-a5c4-4579b08a669a" (UID: "d84a71b6-7503-4150-a5c4-4579b08a669a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.202520 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-scripts" (OuterVolumeSpecName: "scripts") pod "e0f82f51-8a20-402d-8861-6c57a4e5af0d" (UID: "e0f82f51-8a20-402d-8861-6c57a4e5af0d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.246257 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e0f82f51-8a20-402d-8861-6c57a4e5af0d" (UID: "e0f82f51-8a20-402d-8861-6c57a4e5af0d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.254705 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.255036 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jr9d\" (UniqueName: \"kubernetes.io/projected/d84a71b6-7503-4150-a5c4-4579b08a669a-kube-api-access-6jr9d\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.255161 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0f82f51-8a20-402d-8861-6c57a4e5af0d-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.255239 4948 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.255315 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.255370 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5zcb\" (UniqueName: \"kubernetes.io/projected/e0f82f51-8a20-402d-8861-6c57a4e5af0d-kube-api-access-q5zcb\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.255445 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.255519 4948 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e0f82f51-8a20-402d-8861-6c57a4e5af0d-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.255598 4948 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.255854 4948 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.280048 4948 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.292785 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d84a71b6-7503-4150-a5c4-4579b08a669a" (UID: "d84a71b6-7503-4150-a5c4-4579b08a669a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.306431 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-config-data" (OuterVolumeSpecName: "config-data") pod "d84a71b6-7503-4150-a5c4-4579b08a669a" (UID: "d84a71b6-7503-4150-a5c4-4579b08a669a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.322279 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-config-data" (OuterVolumeSpecName: "config-data") pod "e0f82f51-8a20-402d-8861-6c57a4e5af0d" (UID: "e0f82f51-8a20-402d-8861-6c57a4e5af0d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.358490 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.358560 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.358573 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.358583 4948 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.586404 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v29fx" event={"ID":"d84a71b6-7503-4150-a5c4-4579b08a669a","Type":"ContainerDied","Data":"21944c755e2a58a194f7e12be68919b81d0c021ee6380fb185cda3cec0fcee37"} Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.586704 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="21944c755e2a58a194f7e12be68919b81d0c021ee6380fb185cda3cec0fcee37" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.586669 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.601254 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e0f82f51-8a20-402d-8861-6c57a4e5af0d","Type":"ContainerDied","Data":"70cd0cac387b5f659d055b8820f12fcea594407eebec35ab497b573fb1f4db5a"} Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.601308 4948 scope.go:117] "RemoveContainer" containerID="1a5b270055c8b2b5a29d54a228d2fd2e9bdac72aaebc130807e4057de5afdb1a" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.601442 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.620826 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g27hk" event={"ID":"d6729921-0a12-46a0-824f-810df88ae426","Type":"ContainerStarted","Data":"6cd76a3b8281c950d8cac3fc54816ec3fe45fb413dc79249c79a1f2aecbbd206"} Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.626167 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d46988dcb-rn6qs" event={"ID":"8b1ee525-1a6e-461d-aa4d-178ab601ec9d","Type":"ContainerStarted","Data":"6a2830020d67b39ba1edbe811de2ebbd05b2b8a02d0578d95553d45d985544af"} Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.626204 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d46988dcb-rn6qs" event={"ID":"8b1ee525-1a6e-461d-aa4d-178ab601ec9d","Type":"ContainerStarted","Data":"26cbea34a0a499d9db14acd1bf75d24dc0d0a677df528028730bacad671ebd26"} Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.627134 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.627176 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.644437 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hh5s2" event={"ID":"426b37f9-4acb-4d28-a55c-3fcde2600002","Type":"ContainerStarted","Data":"3d2681901d856e71e99e2f5dd2205106a0fb9176c61fcdad6f85c78cea4118e6"} Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.644490 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.665905 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.679892 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.682457 4948 scope.go:117] "RemoveContainer" containerID="7020629c426e657a58b5859f74b5d55f5c4fb964822922a0eb88d95415abe7d8" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.733712 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:22:50 crc kubenswrapper[4948]: E0220 08:22:50.734190 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0f82f51-8a20-402d-8861-6c57a4e5af0d" containerName="glance-log" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.734208 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0f82f51-8a20-402d-8861-6c57a4e5af0d" containerName="glance-log" Feb 20 08:22:50 crc kubenswrapper[4948]: E0220 08:22:50.734246 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0f82f51-8a20-402d-8861-6c57a4e5af0d" containerName="glance-httpd" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.734255 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0f82f51-8a20-402d-8861-6c57a4e5af0d" containerName="glance-httpd" Feb 20 08:22:50 crc kubenswrapper[4948]: E0220 08:22:50.734276 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d84a71b6-7503-4150-a5c4-4579b08a669a" containerName="keystone-bootstrap" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.734283 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d84a71b6-7503-4150-a5c4-4579b08a669a" containerName="keystone-bootstrap" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.734503 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d84a71b6-7503-4150-a5c4-4579b08a669a" containerName="keystone-bootstrap" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.734547 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0f82f51-8a20-402d-8861-6c57a4e5af0d" containerName="glance-httpd" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.734575 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0f82f51-8a20-402d-8861-6c57a4e5af0d" containerName="glance-log" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.735796 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.738323 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.742502 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.748537 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-g27hk" podStartSLOduration=19.425732473 podStartE2EDuration="25.748512013s" podCreationTimestamp="2026-02-20 08:22:25 +0000 UTC" firstStartedPulling="2026-02-20 08:22:43.275291984 +0000 UTC m=+1012.249786794" lastFinishedPulling="2026-02-20 08:22:49.598071514 +0000 UTC m=+1018.572566334" observedRunningTime="2026-02-20 08:22:50.660462804 +0000 UTC m=+1019.634957634" watchObservedRunningTime="2026-02-20 08:22:50.748512013 +0000 UTC m=+1019.723006823" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.763173 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.766915 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-d46988dcb-rn6qs" podStartSLOduration=3.766895298 podStartE2EDuration="3.766895298s" podCreationTimestamp="2026-02-20 08:22:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:22:50.688212221 +0000 UTC m=+1019.662707041" watchObservedRunningTime="2026-02-20 08:22:50.766895298 +0000 UTC m=+1019.741390118" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.779030 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6d4b8df464-wspg2"] Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.780153 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.780631 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hh5s2" podStartSLOduration=16.580718558 podStartE2EDuration="22.780611937s" podCreationTimestamp="2026-02-20 08:22:28 +0000 UTC" firstStartedPulling="2026-02-20 08:22:43.275410777 +0000 UTC m=+1012.249905597" lastFinishedPulling="2026-02-20 08:22:49.475304156 +0000 UTC m=+1018.449798976" observedRunningTime="2026-02-20 08:22:50.719482804 +0000 UTC m=+1019.693977625" watchObservedRunningTime="2026-02-20 08:22:50.780611937 +0000 UTC m=+1019.755106757" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.788506 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-v9rj2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.788689 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.789016 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.789154 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.789308 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.789434 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.861017 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6d4b8df464-wspg2"] Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.874059 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.874150 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-config-data\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.874246 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-credential-keys\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.874276 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d9399b9c-0935-4735-8b08-96d4a29d4ba8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.874301 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-fernet-keys\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.874332 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-combined-ca-bundle\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.874366 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-public-tls-certs\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.874401 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.874455 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-scripts\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.874478 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9399b9c-0935-4735-8b08-96d4a29d4ba8-logs\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.874510 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44bpr\" (UniqueName: \"kubernetes.io/projected/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-kube-api-access-44bpr\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.874541 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdk7s\" (UniqueName: \"kubernetes.io/projected/d9399b9c-0935-4735-8b08-96d4a29d4ba8-kube-api-access-jdk7s\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.874631 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.874653 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-scripts\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.874676 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-config-data\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.874696 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-internal-tls-certs\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.998912 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.998999 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-scripts\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.999045 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-config-data\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.999086 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-internal-tls-certs\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.999168 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.999260 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-config-data\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.999407 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-credential-keys\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.999452 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d9399b9c-0935-4735-8b08-96d4a29d4ba8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.999490 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-fernet-keys\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.999542 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-combined-ca-bundle\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.999606 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-public-tls-certs\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.999669 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.999731 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-scripts\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.999760 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9399b9c-0935-4735-8b08-96d4a29d4ba8-logs\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.999817 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44bpr\" (UniqueName: \"kubernetes.io/projected/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-kube-api-access-44bpr\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.999853 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdk7s\" (UniqueName: \"kubernetes.io/projected/d9399b9c-0935-4735-8b08-96d4a29d4ba8-kube-api-access-jdk7s\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.006035 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-scripts\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.008188 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-internal-tls-certs\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.008718 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-config-data\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.011606 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.013731 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d9399b9c-0935-4735-8b08-96d4a29d4ba8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.015360 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-public-tls-certs\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.020527 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-credential-keys\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.027281 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9399b9c-0935-4735-8b08-96d4a29d4ba8-logs\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.032144 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-combined-ca-bundle\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.035622 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.035633 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-fernet-keys\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.038500 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.039570 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-config-data\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.041985 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44bpr\" (UniqueName: \"kubernetes.io/projected/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-kube-api-access-44bpr\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.045628 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdk7s\" (UniqueName: \"kubernetes.io/projected/d9399b9c-0935-4735-8b08-96d4a29d4ba8-kube-api-access-jdk7s\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.046610 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-scripts\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.077840 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.116525 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.363452 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.691552 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6d4b8df464-wspg2"] Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.760386 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0f82f51-8a20-402d-8861-6c57a4e5af0d" path="/var/lib/kubelet/pods/e0f82f51-8a20-402d-8861-6c57a4e5af0d/volumes" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.812250 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.876364 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-96668"] Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.876589 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-96668" podUID="94b82d55-b467-408f-8eb1-6da4648dde97" containerName="dnsmasq-dns" containerID="cri-o://76e608849a7aba9c1d66b8ca878674048ebd2cf05258476f0e82590946ed89c0" gracePeriod=10 Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.980044 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.980100 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.019854 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.038202 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.136427 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.396304 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.542539 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-ovsdbserver-nb\") pod \"94b82d55-b467-408f-8eb1-6da4648dde97\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.542616 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-ovsdbserver-sb\") pod \"94b82d55-b467-408f-8eb1-6da4648dde97\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.542771 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-dns-swift-storage-0\") pod \"94b82d55-b467-408f-8eb1-6da4648dde97\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.542810 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-dns-svc\") pod \"94b82d55-b467-408f-8eb1-6da4648dde97\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.542864 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-config\") pod \"94b82d55-b467-408f-8eb1-6da4648dde97\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.542887 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqddr\" (UniqueName: \"kubernetes.io/projected/94b82d55-b467-408f-8eb1-6da4648dde97-kube-api-access-rqddr\") pod \"94b82d55-b467-408f-8eb1-6da4648dde97\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.554939 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94b82d55-b467-408f-8eb1-6da4648dde97-kube-api-access-rqddr" (OuterVolumeSpecName: "kube-api-access-rqddr") pod "94b82d55-b467-408f-8eb1-6da4648dde97" (UID: "94b82d55-b467-408f-8eb1-6da4648dde97"). InnerVolumeSpecName "kube-api-access-rqddr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.591136 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "94b82d55-b467-408f-8eb1-6da4648dde97" (UID: "94b82d55-b467-408f-8eb1-6da4648dde97"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.612888 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "94b82d55-b467-408f-8eb1-6da4648dde97" (UID: "94b82d55-b467-408f-8eb1-6da4648dde97"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.616413 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "94b82d55-b467-408f-8eb1-6da4648dde97" (UID: "94b82d55-b467-408f-8eb1-6da4648dde97"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.616531 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "94b82d55-b467-408f-8eb1-6da4648dde97" (UID: "94b82d55-b467-408f-8eb1-6da4648dde97"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.645334 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqddr\" (UniqueName: \"kubernetes.io/projected/94b82d55-b467-408f-8eb1-6da4648dde97-kube-api-access-rqddr\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.645415 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.645429 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.645442 4948 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.645515 4948 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.654497 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-config" (OuterVolumeSpecName: "config") pod "94b82d55-b467-408f-8eb1-6da4648dde97" (UID: "94b82d55-b467-408f-8eb1-6da4648dde97"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.708491 4948 generic.go:334] "Generic (PLEG): container finished" podID="94b82d55-b467-408f-8eb1-6da4648dde97" containerID="76e608849a7aba9c1d66b8ca878674048ebd2cf05258476f0e82590946ed89c0" exitCode=0 Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.708788 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-96668" event={"ID":"94b82d55-b467-408f-8eb1-6da4648dde97","Type":"ContainerDied","Data":"76e608849a7aba9c1d66b8ca878674048ebd2cf05258476f0e82590946ed89c0"} Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.708822 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-96668" event={"ID":"94b82d55-b467-408f-8eb1-6da4648dde97","Type":"ContainerDied","Data":"af3aca3908294e0c1fe45deb6418a29e5d516dc4d54d54ff11acdb6845abd301"} Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.708845 4948 scope.go:117] "RemoveContainer" containerID="76e608849a7aba9c1d66b8ca878674048ebd2cf05258476f0e82590946ed89c0" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.709014 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.713181 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d9399b9c-0935-4735-8b08-96d4a29d4ba8","Type":"ContainerStarted","Data":"fecc6864756af4b1a18a9eaf7ed69fde086ccbc2685b82cc1f598b9178cb31a0"} Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.724409 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6d4b8df464-wspg2" event={"ID":"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc","Type":"ContainerStarted","Data":"6c85d3af2063b59d7ee0e44b6d6917212c7171f281d6864ee9d371d95e5f223a"} Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.733376 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6d4b8df464-wspg2" event={"ID":"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc","Type":"ContainerStarted","Data":"ad469bae9f6d71112f325de3274ec1b7fb44241d69e1e8658079429caacb8930"} Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.733666 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.734161 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.746796 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.750991 4948 scope.go:117] "RemoveContainer" containerID="bbaf3121e36eabf03661c450060f71b1a26f424ea932363040deddf76d19e0df" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.765901 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-6d4b8df464-wspg2" podStartSLOduration=2.765886707 podStartE2EDuration="2.765886707s" podCreationTimestamp="2026-02-20 08:22:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:22:52.749384369 +0000 UTC m=+1021.723879209" watchObservedRunningTime="2026-02-20 08:22:52.765886707 +0000 UTC m=+1021.740381527" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.783039 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-96668"] Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.793139 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-96668"] Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.814387 4948 scope.go:117] "RemoveContainer" containerID="76e608849a7aba9c1d66b8ca878674048ebd2cf05258476f0e82590946ed89c0" Feb 20 08:22:52 crc kubenswrapper[4948]: E0220 08:22:52.814826 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76e608849a7aba9c1d66b8ca878674048ebd2cf05258476f0e82590946ed89c0\": container with ID starting with 76e608849a7aba9c1d66b8ca878674048ebd2cf05258476f0e82590946ed89c0 not found: ID does not exist" containerID="76e608849a7aba9c1d66b8ca878674048ebd2cf05258476f0e82590946ed89c0" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.814856 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76e608849a7aba9c1d66b8ca878674048ebd2cf05258476f0e82590946ed89c0"} err="failed to get container status \"76e608849a7aba9c1d66b8ca878674048ebd2cf05258476f0e82590946ed89c0\": rpc error: code = NotFound desc = could not find container \"76e608849a7aba9c1d66b8ca878674048ebd2cf05258476f0e82590946ed89c0\": container with ID starting with 76e608849a7aba9c1d66b8ca878674048ebd2cf05258476f0e82590946ed89c0 not found: ID does not exist" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.814875 4948 scope.go:117] "RemoveContainer" containerID="bbaf3121e36eabf03661c450060f71b1a26f424ea932363040deddf76d19e0df" Feb 20 08:22:52 crc kubenswrapper[4948]: E0220 08:22:52.815193 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbaf3121e36eabf03661c450060f71b1a26f424ea932363040deddf76d19e0df\": container with ID starting with bbaf3121e36eabf03661c450060f71b1a26f424ea932363040deddf76d19e0df not found: ID does not exist" containerID="bbaf3121e36eabf03661c450060f71b1a26f424ea932363040deddf76d19e0df" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.815212 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbaf3121e36eabf03661c450060f71b1a26f424ea932363040deddf76d19e0df"} err="failed to get container status \"bbaf3121e36eabf03661c450060f71b1a26f424ea932363040deddf76d19e0df\": rpc error: code = NotFound desc = could not find container \"bbaf3121e36eabf03661c450060f71b1a26f424ea932363040deddf76d19e0df\": container with ID starting with bbaf3121e36eabf03661c450060f71b1a26f424ea932363040deddf76d19e0df not found: ID does not exist" Feb 20 08:22:53 crc kubenswrapper[4948]: I0220 08:22:53.736054 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94b82d55-b467-408f-8eb1-6da4648dde97" path="/var/lib/kubelet/pods/94b82d55-b467-408f-8eb1-6da4648dde97/volumes" Feb 20 08:22:53 crc kubenswrapper[4948]: I0220 08:22:53.739271 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d9399b9c-0935-4735-8b08-96d4a29d4ba8","Type":"ContainerStarted","Data":"656e84496ca041e8a454b07ed67e165b130af1030016564ae4519b8e2e6f11e5"} Feb 20 08:22:53 crc kubenswrapper[4948]: I0220 08:22:53.739559 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d9399b9c-0935-4735-8b08-96d4a29d4ba8","Type":"ContainerStarted","Data":"a66ddb00a7f57dc492a3a8a03a3bce06f45c65381b1e5db4f923e10fec5267a0"} Feb 20 08:22:53 crc kubenswrapper[4948]: I0220 08:22:53.740272 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:53 crc kubenswrapper[4948]: I0220 08:22:53.768647 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.768628852 podStartE2EDuration="3.768628852s" podCreationTimestamp="2026-02-20 08:22:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:22:53.755904787 +0000 UTC m=+1022.730399617" watchObservedRunningTime="2026-02-20 08:22:53.768628852 +0000 UTC m=+1022.743123672" Feb 20 08:22:54 crc kubenswrapper[4948]: I0220 08:22:54.747012 4948 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 20 08:22:54 crc kubenswrapper[4948]: I0220 08:22:54.747040 4948 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 20 08:22:54 crc kubenswrapper[4948]: I0220 08:22:54.779914 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 20 08:22:54 crc kubenswrapper[4948]: I0220 08:22:54.781383 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 20 08:22:55 crc kubenswrapper[4948]: I0220 08:22:55.478886 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-66b4d4cb8-hgmkd" podUID="d1046503-c6d3-49bf-9523-2f897fcb82ba" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Feb 20 08:22:55 crc kubenswrapper[4948]: I0220 08:22:55.646403 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-685c4db87b-rw57h" podUID="7b9a072a-a040-414c-96c7-cf1454148745" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Feb 20 08:22:55 crc kubenswrapper[4948]: I0220 08:22:55.966929 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-g27hk" Feb 20 08:22:55 crc kubenswrapper[4948]: I0220 08:22:55.967466 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-g27hk" Feb 20 08:22:56 crc kubenswrapper[4948]: I0220 08:22:56.013505 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-g27hk" Feb 20 08:22:56 crc kubenswrapper[4948]: I0220 08:22:56.845068 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-g27hk" Feb 20 08:22:57 crc kubenswrapper[4948]: I0220 08:22:57.249714 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g27hk"] Feb 20 08:22:57 crc kubenswrapper[4948]: I0220 08:22:57.814588 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-c7zzv" event={"ID":"f1f108e1-27bb-433e-b7e1-f9cc7f778182","Type":"ContainerStarted","Data":"b6240fac3873ea6c61937ec67473ce610d9e47448d45e4e229e635b6cb80cfa1"} Feb 20 08:22:57 crc kubenswrapper[4948]: I0220 08:22:57.820398 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89b49c50-bae7-47fe-b85f-7a718953933d","Type":"ContainerStarted","Data":"61ecb15080679db6c7523b2b45faf0110c6e8bc826d5ac4d37af1077d531ed07"} Feb 20 08:22:57 crc kubenswrapper[4948]: I0220 08:22:57.843832 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-c7zzv" podStartSLOduration=3.872458231 podStartE2EDuration="43.843800229s" podCreationTimestamp="2026-02-20 08:22:14 +0000 UTC" firstStartedPulling="2026-02-20 08:22:16.603201852 +0000 UTC m=+985.577696682" lastFinishedPulling="2026-02-20 08:22:56.57454386 +0000 UTC m=+1025.549038680" observedRunningTime="2026-02-20 08:22:57.829053574 +0000 UTC m=+1026.803548414" watchObservedRunningTime="2026-02-20 08:22:57.843800229 +0000 UTC m=+1026.818295079" Feb 20 08:22:58 crc kubenswrapper[4948]: I0220 08:22:58.829192 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-g27hk" podUID="d6729921-0a12-46a0-824f-810df88ae426" containerName="registry-server" containerID="cri-o://6cd76a3b8281c950d8cac3fc54816ec3fe45fb413dc79249c79a1f2aecbbd206" gracePeriod=2 Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.307666 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g27hk" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.358651 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hh5s2" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.358701 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hh5s2" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.370184 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2k5c\" (UniqueName: \"kubernetes.io/projected/d6729921-0a12-46a0-824f-810df88ae426-kube-api-access-x2k5c\") pod \"d6729921-0a12-46a0-824f-810df88ae426\" (UID: \"d6729921-0a12-46a0-824f-810df88ae426\") " Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.370291 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6729921-0a12-46a0-824f-810df88ae426-utilities\") pod \"d6729921-0a12-46a0-824f-810df88ae426\" (UID: \"d6729921-0a12-46a0-824f-810df88ae426\") " Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.370355 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6729921-0a12-46a0-824f-810df88ae426-catalog-content\") pod \"d6729921-0a12-46a0-824f-810df88ae426\" (UID: \"d6729921-0a12-46a0-824f-810df88ae426\") " Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.371414 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6729921-0a12-46a0-824f-810df88ae426-utilities" (OuterVolumeSpecName: "utilities") pod "d6729921-0a12-46a0-824f-810df88ae426" (UID: "d6729921-0a12-46a0-824f-810df88ae426"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.376014 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6729921-0a12-46a0-824f-810df88ae426-kube-api-access-x2k5c" (OuterVolumeSpecName: "kube-api-access-x2k5c") pod "d6729921-0a12-46a0-824f-810df88ae426" (UID: "d6729921-0a12-46a0-824f-810df88ae426"). InnerVolumeSpecName "kube-api-access-x2k5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.428320 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6729921-0a12-46a0-824f-810df88ae426-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d6729921-0a12-46a0-824f-810df88ae426" (UID: "d6729921-0a12-46a0-824f-810df88ae426"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.472861 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2k5c\" (UniqueName: \"kubernetes.io/projected/d6729921-0a12-46a0-824f-810df88ae426-kube-api-access-x2k5c\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.472897 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6729921-0a12-46a0-824f-810df88ae426-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.472906 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6729921-0a12-46a0-824f-810df88ae426-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.847407 4948 generic.go:334] "Generic (PLEG): container finished" podID="d6729921-0a12-46a0-824f-810df88ae426" containerID="6cd76a3b8281c950d8cac3fc54816ec3fe45fb413dc79249c79a1f2aecbbd206" exitCode=0 Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.847481 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g27hk" event={"ID":"d6729921-0a12-46a0-824f-810df88ae426","Type":"ContainerDied","Data":"6cd76a3b8281c950d8cac3fc54816ec3fe45fb413dc79249c79a1f2aecbbd206"} Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.847512 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g27hk" event={"ID":"d6729921-0a12-46a0-824f-810df88ae426","Type":"ContainerDied","Data":"b7570c60dc7ea9b2dda7deeb7d6ef87e7350a05f972449d37dfa272d75ce2941"} Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.847532 4948 scope.go:117] "RemoveContainer" containerID="6cd76a3b8281c950d8cac3fc54816ec3fe45fb413dc79249c79a1f2aecbbd206" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.847679 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g27hk" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.854662 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-xt2zq" event={"ID":"f706b741-86f9-49a3-95d5-85ee62eb3668","Type":"ContainerStarted","Data":"5b0b3d5fbb40595687dd65feb8573194de8a92b25615987166c82509c29136f4"} Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.857457 4948 generic.go:334] "Generic (PLEG): container finished" podID="f1f108e1-27bb-433e-b7e1-f9cc7f778182" containerID="b6240fac3873ea6c61937ec67473ce610d9e47448d45e4e229e635b6cb80cfa1" exitCode=0 Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.857501 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-c7zzv" event={"ID":"f1f108e1-27bb-433e-b7e1-f9cc7f778182","Type":"ContainerDied","Data":"b6240fac3873ea6c61937ec67473ce610d9e47448d45e4e229e635b6cb80cfa1"} Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.887172 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g27hk"] Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.888955 4948 scope.go:117] "RemoveContainer" containerID="bdb7ba6b04df20596c79e2a31fb3e44aca83f1668de6adef1942e6b1ba6fa793" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.905123 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-g27hk"] Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.914400 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-xt2zq" podStartSLOduration=3.030407265 podStartE2EDuration="45.91437863s" podCreationTimestamp="2026-02-20 08:22:14 +0000 UTC" firstStartedPulling="2026-02-20 08:22:15.460074934 +0000 UTC m=+984.434569754" lastFinishedPulling="2026-02-20 08:22:58.344046299 +0000 UTC m=+1027.318541119" observedRunningTime="2026-02-20 08:22:59.88648426 +0000 UTC m=+1028.860979080" watchObservedRunningTime="2026-02-20 08:22:59.91437863 +0000 UTC m=+1028.888873460" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.925592 4948 scope.go:117] "RemoveContainer" containerID="c30e42a1011ef0a11a3bd3a8a1babce0958085ece91112de32b9f174e1d1c87a" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.955082 4948 scope.go:117] "RemoveContainer" containerID="6cd76a3b8281c950d8cac3fc54816ec3fe45fb413dc79249c79a1f2aecbbd206" Feb 20 08:22:59 crc kubenswrapper[4948]: E0220 08:22:59.956067 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6cd76a3b8281c950d8cac3fc54816ec3fe45fb413dc79249c79a1f2aecbbd206\": container with ID starting with 6cd76a3b8281c950d8cac3fc54816ec3fe45fb413dc79249c79a1f2aecbbd206 not found: ID does not exist" containerID="6cd76a3b8281c950d8cac3fc54816ec3fe45fb413dc79249c79a1f2aecbbd206" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.956104 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6cd76a3b8281c950d8cac3fc54816ec3fe45fb413dc79249c79a1f2aecbbd206"} err="failed to get container status \"6cd76a3b8281c950d8cac3fc54816ec3fe45fb413dc79249c79a1f2aecbbd206\": rpc error: code = NotFound desc = could not find container \"6cd76a3b8281c950d8cac3fc54816ec3fe45fb413dc79249c79a1f2aecbbd206\": container with ID starting with 6cd76a3b8281c950d8cac3fc54816ec3fe45fb413dc79249c79a1f2aecbbd206 not found: ID does not exist" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.956128 4948 scope.go:117] "RemoveContainer" containerID="bdb7ba6b04df20596c79e2a31fb3e44aca83f1668de6adef1942e6b1ba6fa793" Feb 20 08:22:59 crc kubenswrapper[4948]: E0220 08:22:59.957183 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdb7ba6b04df20596c79e2a31fb3e44aca83f1668de6adef1942e6b1ba6fa793\": container with ID starting with bdb7ba6b04df20596c79e2a31fb3e44aca83f1668de6adef1942e6b1ba6fa793 not found: ID does not exist" containerID="bdb7ba6b04df20596c79e2a31fb3e44aca83f1668de6adef1942e6b1ba6fa793" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.957203 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdb7ba6b04df20596c79e2a31fb3e44aca83f1668de6adef1942e6b1ba6fa793"} err="failed to get container status \"bdb7ba6b04df20596c79e2a31fb3e44aca83f1668de6adef1942e6b1ba6fa793\": rpc error: code = NotFound desc = could not find container \"bdb7ba6b04df20596c79e2a31fb3e44aca83f1668de6adef1942e6b1ba6fa793\": container with ID starting with bdb7ba6b04df20596c79e2a31fb3e44aca83f1668de6adef1942e6b1ba6fa793 not found: ID does not exist" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.957222 4948 scope.go:117] "RemoveContainer" containerID="c30e42a1011ef0a11a3bd3a8a1babce0958085ece91112de32b9f174e1d1c87a" Feb 20 08:22:59 crc kubenswrapper[4948]: E0220 08:22:59.957479 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c30e42a1011ef0a11a3bd3a8a1babce0958085ece91112de32b9f174e1d1c87a\": container with ID starting with c30e42a1011ef0a11a3bd3a8a1babce0958085ece91112de32b9f174e1d1c87a not found: ID does not exist" containerID="c30e42a1011ef0a11a3bd3a8a1babce0958085ece91112de32b9f174e1d1c87a" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.957566 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c30e42a1011ef0a11a3bd3a8a1babce0958085ece91112de32b9f174e1d1c87a"} err="failed to get container status \"c30e42a1011ef0a11a3bd3a8a1babce0958085ece91112de32b9f174e1d1c87a\": rpc error: code = NotFound desc = could not find container \"c30e42a1011ef0a11a3bd3a8a1babce0958085ece91112de32b9f174e1d1c87a\": container with ID starting with c30e42a1011ef0a11a3bd3a8a1babce0958085ece91112de32b9f174e1d1c87a not found: ID does not exist" Feb 20 08:23:00 crc kubenswrapper[4948]: I0220 08:23:00.407305 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-hh5s2" podUID="426b37f9-4acb-4d28-a55c-3fcde2600002" containerName="registry-server" probeResult="failure" output=< Feb 20 08:23:00 crc kubenswrapper[4948]: timeout: failed to connect service ":50051" within 1s Feb 20 08:23:00 crc kubenswrapper[4948]: > Feb 20 08:23:01 crc kubenswrapper[4948]: I0220 08:23:01.364694 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 20 08:23:01 crc kubenswrapper[4948]: I0220 08:23:01.364836 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 20 08:23:01 crc kubenswrapper[4948]: I0220 08:23:01.406322 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 20 08:23:01 crc kubenswrapper[4948]: I0220 08:23:01.419667 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 20 08:23:01 crc kubenswrapper[4948]: I0220 08:23:01.737625 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6729921-0a12-46a0-824f-810df88ae426" path="/var/lib/kubelet/pods/d6729921-0a12-46a0-824f-810df88ae426/volumes" Feb 20 08:23:01 crc kubenswrapper[4948]: I0220 08:23:01.883442 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 20 08:23:01 crc kubenswrapper[4948]: I0220 08:23:01.883498 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 20 08:23:02 crc kubenswrapper[4948]: I0220 08:23:02.889174 4948 generic.go:334] "Generic (PLEG): container finished" podID="f706b741-86f9-49a3-95d5-85ee62eb3668" containerID="5b0b3d5fbb40595687dd65feb8573194de8a92b25615987166c82509c29136f4" exitCode=0 Feb 20 08:23:02 crc kubenswrapper[4948]: I0220 08:23:02.889957 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-xt2zq" event={"ID":"f706b741-86f9-49a3-95d5-85ee62eb3668","Type":"ContainerDied","Data":"5b0b3d5fbb40595687dd65feb8573194de8a92b25615987166c82509c29136f4"} Feb 20 08:23:03 crc kubenswrapper[4948]: I0220 08:23:03.641287 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 20 08:23:03 crc kubenswrapper[4948]: I0220 08:23:03.757400 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 20 08:23:04 crc kubenswrapper[4948]: I0220 08:23:04.883523 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:23:04 crc kubenswrapper[4948]: I0220 08:23:04.950216 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-c7zzv" event={"ID":"f1f108e1-27bb-433e-b7e1-f9cc7f778182","Type":"ContainerDied","Data":"1c7c890826add7e27d4d21f7cb357370f931db1e77090c60ac1c40f2ed6d57cc"} Feb 20 08:23:04 crc kubenswrapper[4948]: I0220 08:23:04.950265 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c7c890826add7e27d4d21f7cb357370f931db1e77090c60ac1c40f2ed6d57cc" Feb 20 08:23:04 crc kubenswrapper[4948]: I0220 08:23:04.972164 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:23:04 crc kubenswrapper[4948]: I0220 08:23:04.972331 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-xt2zq" event={"ID":"f706b741-86f9-49a3-95d5-85ee62eb3668","Type":"ContainerDied","Data":"571bfb34e93031079bb93437b09bfef171c7529235d603237690b075a9933988"} Feb 20 08:23:04 crc kubenswrapper[4948]: I0220 08:23:04.972352 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="571bfb34e93031079bb93437b09bfef171c7529235d603237690b075a9933988" Feb 20 08:23:04 crc kubenswrapper[4948]: I0220 08:23:04.983334 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f706b741-86f9-49a3-95d5-85ee62eb3668-etc-machine-id\") pod \"f706b741-86f9-49a3-95d5-85ee62eb3668\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " Feb 20 08:23:04 crc kubenswrapper[4948]: I0220 08:23:04.983521 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-scripts\") pod \"f706b741-86f9-49a3-95d5-85ee62eb3668\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " Feb 20 08:23:04 crc kubenswrapper[4948]: I0220 08:23:04.983453 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f706b741-86f9-49a3-95d5-85ee62eb3668-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f706b741-86f9-49a3-95d5-85ee62eb3668" (UID: "f706b741-86f9-49a3-95d5-85ee62eb3668"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:23:04 crc kubenswrapper[4948]: I0220 08:23:04.983664 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-config-data\") pod \"f706b741-86f9-49a3-95d5-85ee62eb3668\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " Feb 20 08:23:04 crc kubenswrapper[4948]: I0220 08:23:04.983777 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwtv5\" (UniqueName: \"kubernetes.io/projected/f706b741-86f9-49a3-95d5-85ee62eb3668-kube-api-access-fwtv5\") pod \"f706b741-86f9-49a3-95d5-85ee62eb3668\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " Feb 20 08:23:04 crc kubenswrapper[4948]: I0220 08:23:04.983888 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-db-sync-config-data\") pod \"f706b741-86f9-49a3-95d5-85ee62eb3668\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " Feb 20 08:23:04 crc kubenswrapper[4948]: I0220 08:23:04.983981 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-combined-ca-bundle\") pod \"f706b741-86f9-49a3-95d5-85ee62eb3668\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " Feb 20 08:23:04 crc kubenswrapper[4948]: I0220 08:23:04.985614 4948 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f706b741-86f9-49a3-95d5-85ee62eb3668-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:04 crc kubenswrapper[4948]: I0220 08:23:04.987466 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-scripts" (OuterVolumeSpecName: "scripts") pod "f706b741-86f9-49a3-95d5-85ee62eb3668" (UID: "f706b741-86f9-49a3-95d5-85ee62eb3668"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:04 crc kubenswrapper[4948]: I0220 08:23:04.988164 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f706b741-86f9-49a3-95d5-85ee62eb3668-kube-api-access-fwtv5" (OuterVolumeSpecName: "kube-api-access-fwtv5") pod "f706b741-86f9-49a3-95d5-85ee62eb3668" (UID: "f706b741-86f9-49a3-95d5-85ee62eb3668"). InnerVolumeSpecName "kube-api-access-fwtv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:04 crc kubenswrapper[4948]: I0220 08:23:04.988552 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f706b741-86f9-49a3-95d5-85ee62eb3668" (UID: "f706b741-86f9-49a3-95d5-85ee62eb3668"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.008866 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f706b741-86f9-49a3-95d5-85ee62eb3668" (UID: "f706b741-86f9-49a3-95d5-85ee62eb3668"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.024776 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-config-data" (OuterVolumeSpecName: "config-data") pod "f706b741-86f9-49a3-95d5-85ee62eb3668" (UID: "f706b741-86f9-49a3-95d5-85ee62eb3668"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.042633 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-c7zzv" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.086549 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1f108e1-27bb-433e-b7e1-f9cc7f778182-combined-ca-bundle\") pod \"f1f108e1-27bb-433e-b7e1-f9cc7f778182\" (UID: \"f1f108e1-27bb-433e-b7e1-f9cc7f778182\") " Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.086676 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f1f108e1-27bb-433e-b7e1-f9cc7f778182-db-sync-config-data\") pod \"f1f108e1-27bb-433e-b7e1-f9cc7f778182\" (UID: \"f1f108e1-27bb-433e-b7e1-f9cc7f778182\") " Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.086777 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmlgz\" (UniqueName: \"kubernetes.io/projected/f1f108e1-27bb-433e-b7e1-f9cc7f778182-kube-api-access-gmlgz\") pod \"f1f108e1-27bb-433e-b7e1-f9cc7f778182\" (UID: \"f1f108e1-27bb-433e-b7e1-f9cc7f778182\") " Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.087337 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.087360 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.087867 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwtv5\" (UniqueName: \"kubernetes.io/projected/f706b741-86f9-49a3-95d5-85ee62eb3668-kube-api-access-fwtv5\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.087884 4948 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.087896 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.092034 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1f108e1-27bb-433e-b7e1-f9cc7f778182-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f1f108e1-27bb-433e-b7e1-f9cc7f778182" (UID: "f1f108e1-27bb-433e-b7e1-f9cc7f778182"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.093743 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1f108e1-27bb-433e-b7e1-f9cc7f778182-kube-api-access-gmlgz" (OuterVolumeSpecName: "kube-api-access-gmlgz") pod "f1f108e1-27bb-433e-b7e1-f9cc7f778182" (UID: "f1f108e1-27bb-433e-b7e1-f9cc7f778182"). InnerVolumeSpecName "kube-api-access-gmlgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.138901 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1f108e1-27bb-433e-b7e1-f9cc7f778182-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f1f108e1-27bb-433e-b7e1-f9cc7f778182" (UID: "f1f108e1-27bb-433e-b7e1-f9cc7f778182"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.189324 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1f108e1-27bb-433e-b7e1-f9cc7f778182-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.189359 4948 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f1f108e1-27bb-433e-b7e1-f9cc7f778182-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.189372 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmlgz\" (UniqueName: \"kubernetes.io/projected/f1f108e1-27bb-433e-b7e1-f9cc7f778182-kube-api-access-gmlgz\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.221098 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 20 08:23:05 crc kubenswrapper[4948]: E0220 08:23:05.221699 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1f108e1-27bb-433e-b7e1-f9cc7f778182" containerName="barbican-db-sync" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.221804 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1f108e1-27bb-433e-b7e1-f9cc7f778182" containerName="barbican-db-sync" Feb 20 08:23:05 crc kubenswrapper[4948]: E0220 08:23:05.221867 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94b82d55-b467-408f-8eb1-6da4648dde97" containerName="init" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.221918 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="94b82d55-b467-408f-8eb1-6da4648dde97" containerName="init" Feb 20 08:23:05 crc kubenswrapper[4948]: E0220 08:23:05.221996 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94b82d55-b467-408f-8eb1-6da4648dde97" containerName="dnsmasq-dns" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.222059 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="94b82d55-b467-408f-8eb1-6da4648dde97" containerName="dnsmasq-dns" Feb 20 08:23:05 crc kubenswrapper[4948]: E0220 08:23:05.222127 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6729921-0a12-46a0-824f-810df88ae426" containerName="extract-content" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.222181 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6729921-0a12-46a0-824f-810df88ae426" containerName="extract-content" Feb 20 08:23:05 crc kubenswrapper[4948]: E0220 08:23:05.222242 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6729921-0a12-46a0-824f-810df88ae426" containerName="extract-utilities" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.222290 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6729921-0a12-46a0-824f-810df88ae426" containerName="extract-utilities" Feb 20 08:23:05 crc kubenswrapper[4948]: E0220 08:23:05.222345 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f706b741-86f9-49a3-95d5-85ee62eb3668" containerName="cinder-db-sync" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.222413 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f706b741-86f9-49a3-95d5-85ee62eb3668" containerName="cinder-db-sync" Feb 20 08:23:05 crc kubenswrapper[4948]: E0220 08:23:05.222482 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6729921-0a12-46a0-824f-810df88ae426" containerName="registry-server" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.222534 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6729921-0a12-46a0-824f-810df88ae426" containerName="registry-server" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.222732 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="94b82d55-b467-408f-8eb1-6da4648dde97" containerName="dnsmasq-dns" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.222790 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6729921-0a12-46a0-824f-810df88ae426" containerName="registry-server" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.222843 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1f108e1-27bb-433e-b7e1-f9cc7f778182" containerName="barbican-db-sync" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.222910 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f706b741-86f9-49a3-95d5-85ee62eb3668" containerName="cinder-db-sync" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.223808 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.229362 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.243121 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.267016 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b895b5785-97gsc"] Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.268481 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.337507 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b895b5785-97gsc"] Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.391963 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.392043 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-config\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.392067 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-dns-svc\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.392092 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-ovsdbserver-sb\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.392115 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vllr\" (UniqueName: \"kubernetes.io/projected/75294aee-c7f3-4b0b-82fc-df058490b90f-kube-api-access-8vllr\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.392156 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-ovsdbserver-nb\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.392176 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.392230 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-config-data\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.392255 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6wdv\" (UniqueName: \"kubernetes.io/projected/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-kube-api-access-c6wdv\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.392327 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-dns-swift-storage-0\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.392365 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/75294aee-c7f3-4b0b-82fc-df058490b90f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.392386 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-scripts\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.478060 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-66b4d4cb8-hgmkd" podUID="d1046503-c6d3-49bf-9523-2f897fcb82ba" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.494239 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-ovsdbserver-nb\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.494290 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.494328 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-config-data\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.494355 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6wdv\" (UniqueName: \"kubernetes.io/projected/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-kube-api-access-c6wdv\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.494382 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-dns-swift-storage-0\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.494407 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/75294aee-c7f3-4b0b-82fc-df058490b90f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.494422 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-scripts\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.494467 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.494499 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-config\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.494523 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-dns-svc\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.494548 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-ovsdbserver-sb\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.494567 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vllr\" (UniqueName: \"kubernetes.io/projected/75294aee-c7f3-4b0b-82fc-df058490b90f-kube-api-access-8vllr\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.494836 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/75294aee-c7f3-4b0b-82fc-df058490b90f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.496141 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-dns-svc\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.496192 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-config\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.496282 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-ovsdbserver-nb\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.497796 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-dns-swift-storage-0\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.499242 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-ovsdbserver-sb\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.500480 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-scripts\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.502495 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.504805 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-config-data\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.508471 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.517606 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vllr\" (UniqueName: \"kubernetes.io/projected/75294aee-c7f3-4b0b-82fc-df058490b90f-kube-api-access-8vllr\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.521893 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6wdv\" (UniqueName: \"kubernetes.io/projected/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-kube-api-access-c6wdv\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.561827 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.563634 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.565407 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.571082 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.573045 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.623351 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.715820 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xq6j4\" (UniqueName: \"kubernetes.io/projected/1ff6993b-2bde-429d-b0c7-cdea51f733a2-kube-api-access-xq6j4\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.715888 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-config-data-custom\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.715949 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-config-data\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.715984 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.716005 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1ff6993b-2bde-429d-b0c7-cdea51f733a2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.716037 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ff6993b-2bde-429d-b0c7-cdea51f733a2-logs\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.716062 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-scripts\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.817789 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xq6j4\" (UniqueName: \"kubernetes.io/projected/1ff6993b-2bde-429d-b0c7-cdea51f733a2-kube-api-access-xq6j4\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.817865 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-config-data-custom\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.817941 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-config-data\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.817962 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.817994 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1ff6993b-2bde-429d-b0c7-cdea51f733a2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.818029 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ff6993b-2bde-429d-b0c7-cdea51f733a2-logs\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.818046 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-scripts\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.820213 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1ff6993b-2bde-429d-b0c7-cdea51f733a2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.821909 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ff6993b-2bde-429d-b0c7-cdea51f733a2-logs\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.825073 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-scripts\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.831901 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-config-data-custom\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.833162 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-config-data\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.834674 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xq6j4\" (UniqueName: \"kubernetes.io/projected/1ff6993b-2bde-429d-b0c7-cdea51f733a2-kube-api-access-xq6j4\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.842160 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.887683 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.948473 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b895b5785-97gsc"] Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.000205 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b895b5785-97gsc" event={"ID":"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e","Type":"ContainerStarted","Data":"a73968f29763e85cb2562473f92370ee9d88e9ecbb9c1dd9bc348a65b2728c01"} Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.072435 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-c7zzv" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.072519 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89b49c50-bae7-47fe-b85f-7a718953933d","Type":"ContainerStarted","Data":"27fa5fd7b38b4d8b359bee17f753923040dddd64a704b2cd2a15217dcb4497e9"} Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.080345 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="89b49c50-bae7-47fe-b85f-7a718953933d" containerName="ceilometer-central-agent" containerID="cri-o://f21ea8eddb7fd63986bb5357a809bc7c94ba2e30ea5976b9795bdbd7d103ba56" gracePeriod=30 Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.080661 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.080961 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="89b49c50-bae7-47fe-b85f-7a718953933d" containerName="proxy-httpd" containerID="cri-o://27fa5fd7b38b4d8b359bee17f753923040dddd64a704b2cd2a15217dcb4497e9" gracePeriod=30 Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.081041 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="89b49c50-bae7-47fe-b85f-7a718953933d" containerName="sg-core" containerID="cri-o://61ecb15080679db6c7523b2b45faf0110c6e8bc826d5ac4d37af1077d531ed07" gracePeriod=30 Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.081104 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="89b49c50-bae7-47fe-b85f-7a718953933d" containerName="ceilometer-notification-agent" containerID="cri-o://68df1bb1b13a5e62c745d68de5298b184c0cdcaa037a022eedeb800b70856493" gracePeriod=30 Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.089331 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 20 08:23:06 crc kubenswrapper[4948]: W0220 08:23:06.125081 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75294aee_c7f3_4b0b_82fc_df058490b90f.slice/crio-3927e56027e8aee8bf743a6fd812a761cbedf51a0e339d17550adee912677f5f WatchSource:0}: Error finding container 3927e56027e8aee8bf743a6fd812a761cbedf51a0e339d17550adee912677f5f: Status 404 returned error can't find the container with id 3927e56027e8aee8bf743a6fd812a761cbedf51a0e339d17550adee912677f5f Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.142962 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.00916312 podStartE2EDuration="52.142938327s" podCreationTimestamp="2026-02-20 08:22:14 +0000 UTC" firstStartedPulling="2026-02-20 08:22:15.726795425 +0000 UTC m=+984.701290245" lastFinishedPulling="2026-02-20 08:23:04.860570622 +0000 UTC m=+1033.835065452" observedRunningTime="2026-02-20 08:23:06.106921766 +0000 UTC m=+1035.081416596" watchObservedRunningTime="2026-02-20 08:23:06.142938327 +0000 UTC m=+1035.117433147" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.249026 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-76d486c65f-w9lc6"] Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.251097 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-76d486c65f-w9lc6" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.253826 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-8rh2b" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.254467 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.254779 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.266266 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-6ffd468-xgk27"] Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.268272 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6ffd468-xgk27" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.271296 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.281233 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-76d486c65f-w9lc6"] Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.317787 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6ffd468-xgk27"] Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.340913 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/292d8e1e-de3b-4259-afef-c19e2c21187c-combined-ca-bundle\") pod \"barbican-keystone-listener-6ffd468-xgk27\" (UID: \"292d8e1e-de3b-4259-afef-c19e2c21187c\") " pod="openstack/barbican-keystone-listener-6ffd468-xgk27" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.340962 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/292d8e1e-de3b-4259-afef-c19e2c21187c-config-data-custom\") pod \"barbican-keystone-listener-6ffd468-xgk27\" (UID: \"292d8e1e-de3b-4259-afef-c19e2c21187c\") " pod="openstack/barbican-keystone-listener-6ffd468-xgk27" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.341031 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4b356058-d957-4fa5-a106-3fe5052c8539-config-data-custom\") pod \"barbican-worker-76d486c65f-w9lc6\" (UID: \"4b356058-d957-4fa5-a106-3fe5052c8539\") " pod="openstack/barbican-worker-76d486c65f-w9lc6" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.341060 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/292d8e1e-de3b-4259-afef-c19e2c21187c-config-data\") pod \"barbican-keystone-listener-6ffd468-xgk27\" (UID: \"292d8e1e-de3b-4259-afef-c19e2c21187c\") " pod="openstack/barbican-keystone-listener-6ffd468-xgk27" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.341092 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/292d8e1e-de3b-4259-afef-c19e2c21187c-logs\") pod \"barbican-keystone-listener-6ffd468-xgk27\" (UID: \"292d8e1e-de3b-4259-afef-c19e2c21187c\") " pod="openstack/barbican-keystone-listener-6ffd468-xgk27" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.341117 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vz4wb\" (UniqueName: \"kubernetes.io/projected/292d8e1e-de3b-4259-afef-c19e2c21187c-kube-api-access-vz4wb\") pod \"barbican-keystone-listener-6ffd468-xgk27\" (UID: \"292d8e1e-de3b-4259-afef-c19e2c21187c\") " pod="openstack/barbican-keystone-listener-6ffd468-xgk27" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.341140 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b356058-d957-4fa5-a106-3fe5052c8539-combined-ca-bundle\") pod \"barbican-worker-76d486c65f-w9lc6\" (UID: \"4b356058-d957-4fa5-a106-3fe5052c8539\") " pod="openstack/barbican-worker-76d486c65f-w9lc6" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.341160 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b356058-d957-4fa5-a106-3fe5052c8539-logs\") pod \"barbican-worker-76d486c65f-w9lc6\" (UID: \"4b356058-d957-4fa5-a106-3fe5052c8539\") " pod="openstack/barbican-worker-76d486c65f-w9lc6" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.341183 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65fwq\" (UniqueName: \"kubernetes.io/projected/4b356058-d957-4fa5-a106-3fe5052c8539-kube-api-access-65fwq\") pod \"barbican-worker-76d486c65f-w9lc6\" (UID: \"4b356058-d957-4fa5-a106-3fe5052c8539\") " pod="openstack/barbican-worker-76d486c65f-w9lc6" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.341206 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b356058-d957-4fa5-a106-3fe5052c8539-config-data\") pod \"barbican-worker-76d486c65f-w9lc6\" (UID: \"4b356058-d957-4fa5-a106-3fe5052c8539\") " pod="openstack/barbican-worker-76d486c65f-w9lc6" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.346121 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b895b5785-97gsc"] Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.361286 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-5dgw7"] Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.368446 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.382180 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-5dgw7"] Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.421042 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.446518 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b356058-d957-4fa5-a106-3fe5052c8539-logs\") pod \"barbican-worker-76d486c65f-w9lc6\" (UID: \"4b356058-d957-4fa5-a106-3fe5052c8539\") " pod="openstack/barbican-worker-76d486c65f-w9lc6" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.446569 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65fwq\" (UniqueName: \"kubernetes.io/projected/4b356058-d957-4fa5-a106-3fe5052c8539-kube-api-access-65fwq\") pod \"barbican-worker-76d486c65f-w9lc6\" (UID: \"4b356058-d957-4fa5-a106-3fe5052c8539\") " pod="openstack/barbican-worker-76d486c65f-w9lc6" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.446596 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b356058-d957-4fa5-a106-3fe5052c8539-config-data\") pod \"barbican-worker-76d486c65f-w9lc6\" (UID: \"4b356058-d957-4fa5-a106-3fe5052c8539\") " pod="openstack/barbican-worker-76d486c65f-w9lc6" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.446627 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.446649 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-config\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.446696 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/292d8e1e-de3b-4259-afef-c19e2c21187c-combined-ca-bundle\") pod \"barbican-keystone-listener-6ffd468-xgk27\" (UID: \"292d8e1e-de3b-4259-afef-c19e2c21187c\") " pod="openstack/barbican-keystone-listener-6ffd468-xgk27" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.446721 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/292d8e1e-de3b-4259-afef-c19e2c21187c-config-data-custom\") pod \"barbican-keystone-listener-6ffd468-xgk27\" (UID: \"292d8e1e-de3b-4259-afef-c19e2c21187c\") " pod="openstack/barbican-keystone-listener-6ffd468-xgk27" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.446745 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2cgj\" (UniqueName: \"kubernetes.io/projected/63124eef-9456-44d1-a8ba-cc8297af3e9b-kube-api-access-l2cgj\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.446770 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.446795 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4b356058-d957-4fa5-a106-3fe5052c8539-config-data-custom\") pod \"barbican-worker-76d486c65f-w9lc6\" (UID: \"4b356058-d957-4fa5-a106-3fe5052c8539\") " pod="openstack/barbican-worker-76d486c65f-w9lc6" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.446829 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/292d8e1e-de3b-4259-afef-c19e2c21187c-config-data\") pod \"barbican-keystone-listener-6ffd468-xgk27\" (UID: \"292d8e1e-de3b-4259-afef-c19e2c21187c\") " pod="openstack/barbican-keystone-listener-6ffd468-xgk27" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.446852 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.446874 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/292d8e1e-de3b-4259-afef-c19e2c21187c-logs\") pod \"barbican-keystone-listener-6ffd468-xgk27\" (UID: \"292d8e1e-de3b-4259-afef-c19e2c21187c\") " pod="openstack/barbican-keystone-listener-6ffd468-xgk27" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.446900 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.446929 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vz4wb\" (UniqueName: \"kubernetes.io/projected/292d8e1e-de3b-4259-afef-c19e2c21187c-kube-api-access-vz4wb\") pod \"barbican-keystone-listener-6ffd468-xgk27\" (UID: \"292d8e1e-de3b-4259-afef-c19e2c21187c\") " pod="openstack/barbican-keystone-listener-6ffd468-xgk27" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.446961 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b356058-d957-4fa5-a106-3fe5052c8539-combined-ca-bundle\") pod \"barbican-worker-76d486c65f-w9lc6\" (UID: \"4b356058-d957-4fa5-a106-3fe5052c8539\") " pod="openstack/barbican-worker-76d486c65f-w9lc6" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.448139 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b356058-d957-4fa5-a106-3fe5052c8539-logs\") pod \"barbican-worker-76d486c65f-w9lc6\" (UID: \"4b356058-d957-4fa5-a106-3fe5052c8539\") " pod="openstack/barbican-worker-76d486c65f-w9lc6" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.453480 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/292d8e1e-de3b-4259-afef-c19e2c21187c-logs\") pod \"barbican-keystone-listener-6ffd468-xgk27\" (UID: \"292d8e1e-de3b-4259-afef-c19e2c21187c\") " pod="openstack/barbican-keystone-listener-6ffd468-xgk27" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.459184 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b356058-d957-4fa5-a106-3fe5052c8539-combined-ca-bundle\") pod \"barbican-worker-76d486c65f-w9lc6\" (UID: \"4b356058-d957-4fa5-a106-3fe5052c8539\") " pod="openstack/barbican-worker-76d486c65f-w9lc6" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.459183 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/292d8e1e-de3b-4259-afef-c19e2c21187c-config-data-custom\") pod \"barbican-keystone-listener-6ffd468-xgk27\" (UID: \"292d8e1e-de3b-4259-afef-c19e2c21187c\") " pod="openstack/barbican-keystone-listener-6ffd468-xgk27" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.461370 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/292d8e1e-de3b-4259-afef-c19e2c21187c-combined-ca-bundle\") pod \"barbican-keystone-listener-6ffd468-xgk27\" (UID: \"292d8e1e-de3b-4259-afef-c19e2c21187c\") " pod="openstack/barbican-keystone-listener-6ffd468-xgk27" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.461891 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4b356058-d957-4fa5-a106-3fe5052c8539-config-data-custom\") pod \"barbican-worker-76d486c65f-w9lc6\" (UID: \"4b356058-d957-4fa5-a106-3fe5052c8539\") " pod="openstack/barbican-worker-76d486c65f-w9lc6" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.462859 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b356058-d957-4fa5-a106-3fe5052c8539-config-data\") pod \"barbican-worker-76d486c65f-w9lc6\" (UID: \"4b356058-d957-4fa5-a106-3fe5052c8539\") " pod="openstack/barbican-worker-76d486c65f-w9lc6" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.468626 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65fwq\" (UniqueName: \"kubernetes.io/projected/4b356058-d957-4fa5-a106-3fe5052c8539-kube-api-access-65fwq\") pod \"barbican-worker-76d486c65f-w9lc6\" (UID: \"4b356058-d957-4fa5-a106-3fe5052c8539\") " pod="openstack/barbican-worker-76d486c65f-w9lc6" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.468984 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/292d8e1e-de3b-4259-afef-c19e2c21187c-config-data\") pod \"barbican-keystone-listener-6ffd468-xgk27\" (UID: \"292d8e1e-de3b-4259-afef-c19e2c21187c\") " pod="openstack/barbican-keystone-listener-6ffd468-xgk27" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.475288 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vz4wb\" (UniqueName: \"kubernetes.io/projected/292d8e1e-de3b-4259-afef-c19e2c21187c-kube-api-access-vz4wb\") pod \"barbican-keystone-listener-6ffd468-xgk27\" (UID: \"292d8e1e-de3b-4259-afef-c19e2c21187c\") " pod="openstack/barbican-keystone-listener-6ffd468-xgk27" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.499326 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-65487cc4d6-7gdvr"] Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.503673 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.507396 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.520613 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-65487cc4d6-7gdvr"] Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.549375 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.549423 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-config\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.549486 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2cgj\" (UniqueName: \"kubernetes.io/projected/63124eef-9456-44d1-a8ba-cc8297af3e9b-kube-api-access-l2cgj\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.549520 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.549582 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.549643 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.551742 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.552287 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.552868 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-config\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.554268 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.555374 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.571403 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2cgj\" (UniqueName: \"kubernetes.io/projected/63124eef-9456-44d1-a8ba-cc8297af3e9b-kube-api-access-l2cgj\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.597895 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-76d486c65f-w9lc6" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.613670 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6ffd468-xgk27" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.650749 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ld59l\" (UniqueName: \"kubernetes.io/projected/6a669dd0-d287-4d07-9eec-aea5ed80c791-kube-api-access-ld59l\") pod \"barbican-api-65487cc4d6-7gdvr\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.650801 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-config-data-custom\") pod \"barbican-api-65487cc4d6-7gdvr\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.650822 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-combined-ca-bundle\") pod \"barbican-api-65487cc4d6-7gdvr\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.651248 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a669dd0-d287-4d07-9eec-aea5ed80c791-logs\") pod \"barbican-api-65487cc4d6-7gdvr\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.651372 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-config-data\") pod \"barbican-api-65487cc4d6-7gdvr\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.691826 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.760992 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a669dd0-d287-4d07-9eec-aea5ed80c791-logs\") pod \"barbican-api-65487cc4d6-7gdvr\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.761044 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-config-data\") pod \"barbican-api-65487cc4d6-7gdvr\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.761075 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ld59l\" (UniqueName: \"kubernetes.io/projected/6a669dd0-d287-4d07-9eec-aea5ed80c791-kube-api-access-ld59l\") pod \"barbican-api-65487cc4d6-7gdvr\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.761108 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-config-data-custom\") pod \"barbican-api-65487cc4d6-7gdvr\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.761127 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-combined-ca-bundle\") pod \"barbican-api-65487cc4d6-7gdvr\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.762350 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a669dd0-d287-4d07-9eec-aea5ed80c791-logs\") pod \"barbican-api-65487cc4d6-7gdvr\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.767192 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-config-data-custom\") pod \"barbican-api-65487cc4d6-7gdvr\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.767391 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-combined-ca-bundle\") pod \"barbican-api-65487cc4d6-7gdvr\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.768570 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-config-data\") pod \"barbican-api-65487cc4d6-7gdvr\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.785232 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ld59l\" (UniqueName: \"kubernetes.io/projected/6a669dd0-d287-4d07-9eec-aea5ed80c791-kube-api-access-ld59l\") pod \"barbican-api-65487cc4d6-7gdvr\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.828051 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.059503 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-76d486c65f-w9lc6"] Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.101983 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-76d486c65f-w9lc6" event={"ID":"4b356058-d957-4fa5-a106-3fe5052c8539","Type":"ContainerStarted","Data":"6d267613f41bda4b9e00a369fc1b5681c20f7663cecb7ffde0195009376088c3"} Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.109793 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1ff6993b-2bde-429d-b0c7-cdea51f733a2","Type":"ContainerStarted","Data":"b0f8def01191a21afa9e066696a67301fbaaf496e33dd51ae806612c580bc882"} Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.112887 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"75294aee-c7f3-4b0b-82fc-df058490b90f","Type":"ContainerStarted","Data":"3927e56027e8aee8bf743a6fd812a761cbedf51a0e339d17550adee912677f5f"} Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.121808 4948 generic.go:334] "Generic (PLEG): container finished" podID="4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e" containerID="a054eac97e88cce1aa9882153236f28817412a343a3608ef402b154e16e8b759" exitCode=0 Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.121927 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b895b5785-97gsc" event={"ID":"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e","Type":"ContainerDied","Data":"a054eac97e88cce1aa9882153236f28817412a343a3608ef402b154e16e8b759"} Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.134758 4948 generic.go:334] "Generic (PLEG): container finished" podID="89b49c50-bae7-47fe-b85f-7a718953933d" containerID="27fa5fd7b38b4d8b359bee17f753923040dddd64a704b2cd2a15217dcb4497e9" exitCode=0 Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.134789 4948 generic.go:334] "Generic (PLEG): container finished" podID="89b49c50-bae7-47fe-b85f-7a718953933d" containerID="61ecb15080679db6c7523b2b45faf0110c6e8bc826d5ac4d37af1077d531ed07" exitCode=2 Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.134797 4948 generic.go:334] "Generic (PLEG): container finished" podID="89b49c50-bae7-47fe-b85f-7a718953933d" containerID="f21ea8eddb7fd63986bb5357a809bc7c94ba2e30ea5976b9795bdbd7d103ba56" exitCode=0 Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.134815 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89b49c50-bae7-47fe-b85f-7a718953933d","Type":"ContainerDied","Data":"27fa5fd7b38b4d8b359bee17f753923040dddd64a704b2cd2a15217dcb4497e9"} Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.134841 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89b49c50-bae7-47fe-b85f-7a718953933d","Type":"ContainerDied","Data":"61ecb15080679db6c7523b2b45faf0110c6e8bc826d5ac4d37af1077d531ed07"} Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.134852 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89b49c50-bae7-47fe-b85f-7a718953933d","Type":"ContainerDied","Data":"f21ea8eddb7fd63986bb5357a809bc7c94ba2e30ea5976b9795bdbd7d103ba56"} Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.167180 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6ffd468-xgk27"] Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.362083 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-5dgw7"] Feb 20 08:23:07 crc kubenswrapper[4948]: W0220 08:23:07.369658 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63124eef_9456_44d1_a8ba_cc8297af3e9b.slice/crio-b02b398d101ad6f54b2deb0825184d401d1e5b0ad4e38257f4120a442c5a5509 WatchSource:0}: Error finding container b02b398d101ad6f54b2deb0825184d401d1e5b0ad4e38257f4120a442c5a5509: Status 404 returned error can't find the container with id b02b398d101ad6f54b2deb0825184d401d1e5b0ad4e38257f4120a442c5a5509 Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.412307 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-65487cc4d6-7gdvr"] Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.433301 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.584173 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-dns-swift-storage-0\") pod \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.584272 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6wdv\" (UniqueName: \"kubernetes.io/projected/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-kube-api-access-c6wdv\") pod \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.584317 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-dns-svc\") pod \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.584348 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-config\") pod \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.584389 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-ovsdbserver-sb\") pod \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.584416 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-ovsdbserver-nb\") pod \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.596175 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-kube-api-access-c6wdv" (OuterVolumeSpecName: "kube-api-access-c6wdv") pod "4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e" (UID: "4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e"). InnerVolumeSpecName "kube-api-access-c6wdv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.625576 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e" (UID: "4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.635254 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e" (UID: "4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.637472 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-config" (OuterVolumeSpecName: "config") pod "4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e" (UID: "4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.638838 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e" (UID: "4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.639942 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e" (UID: "4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.688433 4948 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.688513 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6wdv\" (UniqueName: \"kubernetes.io/projected/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-kube-api-access-c6wdv\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.688536 4948 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.688550 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.688600 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.688613 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.888225 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 20 08:23:08 crc kubenswrapper[4948]: I0220 08:23:08.049995 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:23:08 crc kubenswrapper[4948]: I0220 08:23:08.144909 4948 generic.go:334] "Generic (PLEG): container finished" podID="63124eef-9456-44d1-a8ba-cc8297af3e9b" containerID="56e628ea61223b6bdd0b21738d110aaaa403a55ddd583614f61bc57a6cdb16fd" exitCode=0 Feb 20 08:23:08 crc kubenswrapper[4948]: I0220 08:23:08.145023 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" event={"ID":"63124eef-9456-44d1-a8ba-cc8297af3e9b","Type":"ContainerDied","Data":"56e628ea61223b6bdd0b21738d110aaaa403a55ddd583614f61bc57a6cdb16fd"} Feb 20 08:23:08 crc kubenswrapper[4948]: I0220 08:23:08.145087 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" event={"ID":"63124eef-9456-44d1-a8ba-cc8297af3e9b","Type":"ContainerStarted","Data":"b02b398d101ad6f54b2deb0825184d401d1e5b0ad4e38257f4120a442c5a5509"} Feb 20 08:23:08 crc kubenswrapper[4948]: I0220 08:23:08.149984 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1ff6993b-2bde-429d-b0c7-cdea51f733a2","Type":"ContainerStarted","Data":"2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d"} Feb 20 08:23:08 crc kubenswrapper[4948]: I0220 08:23:08.156378 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6ffd468-xgk27" event={"ID":"292d8e1e-de3b-4259-afef-c19e2c21187c","Type":"ContainerStarted","Data":"f6496764c718a0f648853d16537d4d7261c22048971ad4db61bf7b138d835187"} Feb 20 08:23:08 crc kubenswrapper[4948]: I0220 08:23:08.157769 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"75294aee-c7f3-4b0b-82fc-df058490b90f","Type":"ContainerStarted","Data":"0c8a24ef5e8c2ed45ddfc6908331dbda28dd167039cb8bc7f442c18afef4d964"} Feb 20 08:23:08 crc kubenswrapper[4948]: I0220 08:23:08.161089 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65487cc4d6-7gdvr" event={"ID":"6a669dd0-d287-4d07-9eec-aea5ed80c791","Type":"ContainerStarted","Data":"86e858c136de80a807f997b8991e918a47db690a2c7b9fcc3401caa707f989dd"} Feb 20 08:23:08 crc kubenswrapper[4948]: I0220 08:23:08.161150 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65487cc4d6-7gdvr" event={"ID":"6a669dd0-d287-4d07-9eec-aea5ed80c791","Type":"ContainerStarted","Data":"b3b840325250a810c158ca1f61d60db618feec352dbf68e16aad14e15139327d"} Feb 20 08:23:08 crc kubenswrapper[4948]: I0220 08:23:08.168939 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b895b5785-97gsc" event={"ID":"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e","Type":"ContainerDied","Data":"a73968f29763e85cb2562473f92370ee9d88e9ecbb9c1dd9bc348a65b2728c01"} Feb 20 08:23:08 crc kubenswrapper[4948]: I0220 08:23:08.168996 4948 scope.go:117] "RemoveContainer" containerID="a054eac97e88cce1aa9882153236f28817412a343a3608ef402b154e16e8b759" Feb 20 08:23:08 crc kubenswrapper[4948]: I0220 08:23:08.169146 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:08 crc kubenswrapper[4948]: I0220 08:23:08.388034 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b895b5785-97gsc"] Feb 20 08:23:08 crc kubenswrapper[4948]: I0220 08:23:08.401235 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b895b5785-97gsc"] Feb 20 08:23:09 crc kubenswrapper[4948]: I0220 08:23:09.429695 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hh5s2" Feb 20 08:23:09 crc kubenswrapper[4948]: I0220 08:23:09.526541 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hh5s2" Feb 20 08:23:09 crc kubenswrapper[4948]: I0220 08:23:09.666147 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hh5s2"] Feb 20 08:23:09 crc kubenswrapper[4948]: I0220 08:23:09.738320 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e" path="/var/lib/kubelet/pods/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e/volumes" Feb 20 08:23:09 crc kubenswrapper[4948]: I0220 08:23:09.965399 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.059477 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-66b4d4cb8-hgmkd"] Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.059724 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-66b4d4cb8-hgmkd" podUID="d1046503-c6d3-49bf-9523-2f897fcb82ba" containerName="horizon-log" containerID="cri-o://577865ca69f0d8457d617948e5cac2be0a124e4620edc60df1395130de8a2334" gracePeriod=30 Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.059783 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-66b4d4cb8-hgmkd" podUID="d1046503-c6d3-49bf-9523-2f897fcb82ba" containerName="horizon" containerID="cri-o://a1b60a312d88b84db5225c1aff5290457499b2775b6fffec7ca3246896b28523" gracePeriod=30 Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.190084 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1ff6993b-2bde-429d-b0c7-cdea51f733a2","Type":"ContainerStarted","Data":"0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3"} Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.190170 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.190159 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="1ff6993b-2bde-429d-b0c7-cdea51f733a2" containerName="cinder-api-log" containerID="cri-o://2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d" gracePeriod=30 Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.190218 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="1ff6993b-2bde-429d-b0c7-cdea51f733a2" containerName="cinder-api" containerID="cri-o://0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3" gracePeriod=30 Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.194571 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"75294aee-c7f3-4b0b-82fc-df058490b90f","Type":"ContainerStarted","Data":"c1fbafc4d60f4668a3ce40442dbec31bba82bcb8754ebb88a8aed726953af8e7"} Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.200753 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65487cc4d6-7gdvr" event={"ID":"6a669dd0-d287-4d07-9eec-aea5ed80c791","Type":"ContainerStarted","Data":"ab05c86cbd431ad2cc4a3cef37a6fc481837d1b6073c9a80ed0ee92d983240c8"} Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.200986 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.202801 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" event={"ID":"63124eef-9456-44d1-a8ba-cc8297af3e9b","Type":"ContainerStarted","Data":"724a55a11670e45fe3d2ce7a82832ba87e341cd787eee57c186dd30931fb8402"} Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.212951 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.212936127 podStartE2EDuration="5.212936127s" podCreationTimestamp="2026-02-20 08:23:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:23:10.2065656 +0000 UTC m=+1039.181060430" watchObservedRunningTime="2026-02-20 08:23:10.212936127 +0000 UTC m=+1039.187430947" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.234557 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.472942855 podStartE2EDuration="5.234537932s" podCreationTimestamp="2026-02-20 08:23:05 +0000 UTC" firstStartedPulling="2026-02-20 08:23:06.127056574 +0000 UTC m=+1035.101551394" lastFinishedPulling="2026-02-20 08:23:06.888651651 +0000 UTC m=+1035.863146471" observedRunningTime="2026-02-20 08:23:10.231069406 +0000 UTC m=+1039.205564226" watchObservedRunningTime="2026-02-20 08:23:10.234537932 +0000 UTC m=+1039.209032752" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.257947 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-65487cc4d6-7gdvr" podStartSLOduration=4.25792403 podStartE2EDuration="4.25792403s" podCreationTimestamp="2026-02-20 08:23:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:23:10.247431931 +0000 UTC m=+1039.221926761" watchObservedRunningTime="2026-02-20 08:23:10.25792403 +0000 UTC m=+1039.232418850" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.271674 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" podStartSLOduration=4.27165913 podStartE2EDuration="4.27165913s" podCreationTimestamp="2026-02-20 08:23:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:23:10.268725188 +0000 UTC m=+1039.243220008" watchObservedRunningTime="2026-02-20 08:23:10.27165913 +0000 UTC m=+1039.246153950" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.562881 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.804269 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.870126 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.958949 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-scripts\") pod \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.959364 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xq6j4\" (UniqueName: \"kubernetes.io/projected/1ff6993b-2bde-429d-b0c7-cdea51f733a2-kube-api-access-xq6j4\") pod \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.959385 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1ff6993b-2bde-429d-b0c7-cdea51f733a2-etc-machine-id\") pod \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.959404 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-combined-ca-bundle\") pod \"89b49c50-bae7-47fe-b85f-7a718953933d\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.959475 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77njn\" (UniqueName: \"kubernetes.io/projected/89b49c50-bae7-47fe-b85f-7a718953933d-kube-api-access-77njn\") pod \"89b49c50-bae7-47fe-b85f-7a718953933d\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.959500 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89b49c50-bae7-47fe-b85f-7a718953933d-log-httpd\") pod \"89b49c50-bae7-47fe-b85f-7a718953933d\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.959521 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-scripts\") pod \"89b49c50-bae7-47fe-b85f-7a718953933d\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.959552 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ff6993b-2bde-429d-b0c7-cdea51f733a2-logs\") pod \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.959573 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-config-data\") pod \"89b49c50-bae7-47fe-b85f-7a718953933d\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.959582 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ff6993b-2bde-429d-b0c7-cdea51f733a2-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1ff6993b-2bde-429d-b0c7-cdea51f733a2" (UID: "1ff6993b-2bde-429d-b0c7-cdea51f733a2"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.959607 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-combined-ca-bundle\") pod \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.959734 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-sg-core-conf-yaml\") pod \"89b49c50-bae7-47fe-b85f-7a718953933d\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.959790 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89b49c50-bae7-47fe-b85f-7a718953933d-run-httpd\") pod \"89b49c50-bae7-47fe-b85f-7a718953933d\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.959815 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-config-data\") pod \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.959843 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-config-data-custom\") pod \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.960274 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ff6993b-2bde-429d-b0c7-cdea51f733a2-logs" (OuterVolumeSpecName: "logs") pod "1ff6993b-2bde-429d-b0c7-cdea51f733a2" (UID: "1ff6993b-2bde-429d-b0c7-cdea51f733a2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.960299 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89b49c50-bae7-47fe-b85f-7a718953933d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "89b49c50-bae7-47fe-b85f-7a718953933d" (UID: "89b49c50-bae7-47fe-b85f-7a718953933d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.960735 4948 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89b49c50-bae7-47fe-b85f-7a718953933d-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.960757 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ff6993b-2bde-429d-b0c7-cdea51f733a2-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.960901 4948 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1ff6993b-2bde-429d-b0c7-cdea51f733a2-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.960934 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89b49c50-bae7-47fe-b85f-7a718953933d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "89b49c50-bae7-47fe-b85f-7a718953933d" (UID: "89b49c50-bae7-47fe-b85f-7a718953933d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.963770 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89b49c50-bae7-47fe-b85f-7a718953933d-kube-api-access-77njn" (OuterVolumeSpecName: "kube-api-access-77njn") pod "89b49c50-bae7-47fe-b85f-7a718953933d" (UID: "89b49c50-bae7-47fe-b85f-7a718953933d"). InnerVolumeSpecName "kube-api-access-77njn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.966717 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-scripts" (OuterVolumeSpecName: "scripts") pod "89b49c50-bae7-47fe-b85f-7a718953933d" (UID: "89b49c50-bae7-47fe-b85f-7a718953933d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.969392 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-scripts" (OuterVolumeSpecName: "scripts") pod "1ff6993b-2bde-429d-b0c7-cdea51f733a2" (UID: "1ff6993b-2bde-429d-b0c7-cdea51f733a2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.969417 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ff6993b-2bde-429d-b0c7-cdea51f733a2-kube-api-access-xq6j4" (OuterVolumeSpecName: "kube-api-access-xq6j4") pod "1ff6993b-2bde-429d-b0c7-cdea51f733a2" (UID: "1ff6993b-2bde-429d-b0c7-cdea51f733a2"). InnerVolumeSpecName "kube-api-access-xq6j4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.970223 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1ff6993b-2bde-429d-b0c7-cdea51f733a2" (UID: "1ff6993b-2bde-429d-b0c7-cdea51f733a2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.992370 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1ff6993b-2bde-429d-b0c7-cdea51f733a2" (UID: "1ff6993b-2bde-429d-b0c7-cdea51f733a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.992741 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "89b49c50-bae7-47fe-b85f-7a718953933d" (UID: "89b49c50-bae7-47fe-b85f-7a718953933d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.016496 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-config-data" (OuterVolumeSpecName: "config-data") pod "1ff6993b-2bde-429d-b0c7-cdea51f733a2" (UID: "1ff6993b-2bde-429d-b0c7-cdea51f733a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.038190 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "89b49c50-bae7-47fe-b85f-7a718953933d" (UID: "89b49c50-bae7-47fe-b85f-7a718953933d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.061765 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-config-data" (OuterVolumeSpecName: "config-data") pod "89b49c50-bae7-47fe-b85f-7a718953933d" (UID: "89b49c50-bae7-47fe-b85f-7a718953933d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.062135 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.062160 4948 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.062171 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.062179 4948 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89b49c50-bae7-47fe-b85f-7a718953933d-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.062188 4948 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.062195 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.062205 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xq6j4\" (UniqueName: \"kubernetes.io/projected/1ff6993b-2bde-429d-b0c7-cdea51f733a2-kube-api-access-xq6j4\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.062215 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.062223 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77njn\" (UniqueName: \"kubernetes.io/projected/89b49c50-bae7-47fe-b85f-7a718953933d-kube-api-access-77njn\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.062231 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.062238 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.216274 4948 generic.go:334] "Generic (PLEG): container finished" podID="d1046503-c6d3-49bf-9523-2f897fcb82ba" containerID="a1b60a312d88b84db5225c1aff5290457499b2775b6fffec7ca3246896b28523" exitCode=0 Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.216362 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66b4d4cb8-hgmkd" event={"ID":"d1046503-c6d3-49bf-9523-2f897fcb82ba","Type":"ContainerDied","Data":"a1b60a312d88b84db5225c1aff5290457499b2775b6fffec7ca3246896b28523"} Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.224249 4948 generic.go:334] "Generic (PLEG): container finished" podID="89b49c50-bae7-47fe-b85f-7a718953933d" containerID="68df1bb1b13a5e62c745d68de5298b184c0cdcaa037a022eedeb800b70856493" exitCode=0 Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.224283 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89b49c50-bae7-47fe-b85f-7a718953933d","Type":"ContainerDied","Data":"68df1bb1b13a5e62c745d68de5298b184c0cdcaa037a022eedeb800b70856493"} Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.224336 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89b49c50-bae7-47fe-b85f-7a718953933d","Type":"ContainerDied","Data":"2c29cb12a24e0769e2fc11186d9dff0648d1bcb261fcbab7060faff8fb76085a"} Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.224347 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.224358 4948 scope.go:117] "RemoveContainer" containerID="27fa5fd7b38b4d8b359bee17f753923040dddd64a704b2cd2a15217dcb4497e9" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.227345 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-76d486c65f-w9lc6" event={"ID":"4b356058-d957-4fa5-a106-3fe5052c8539","Type":"ContainerStarted","Data":"429d83e2cd3914efc5db9d90e28d0b14b09e14a603ac3bc8a797688781218235"} Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.227385 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-76d486c65f-w9lc6" event={"ID":"4b356058-d957-4fa5-a106-3fe5052c8539","Type":"ContainerStarted","Data":"97fe5c50373074a6e2530e57e930ec6666df51919433a7f7204067aad9e98819"} Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.231553 4948 generic.go:334] "Generic (PLEG): container finished" podID="1ff6993b-2bde-429d-b0c7-cdea51f733a2" containerID="0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3" exitCode=0 Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.231593 4948 generic.go:334] "Generic (PLEG): container finished" podID="1ff6993b-2bde-429d-b0c7-cdea51f733a2" containerID="2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d" exitCode=143 Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.231639 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1ff6993b-2bde-429d-b0c7-cdea51f733a2","Type":"ContainerDied","Data":"0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3"} Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.231666 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1ff6993b-2bde-429d-b0c7-cdea51f733a2","Type":"ContainerDied","Data":"2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d"} Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.231684 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1ff6993b-2bde-429d-b0c7-cdea51f733a2","Type":"ContainerDied","Data":"b0f8def01191a21afa9e066696a67301fbaaf496e33dd51ae806612c580bc882"} Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.231754 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.251411 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6ffd468-xgk27" event={"ID":"292d8e1e-de3b-4259-afef-c19e2c21187c","Type":"ContainerStarted","Data":"06bcf0b4075319148a17c9d445b9a694bf41ad80dfe19d1b6d5b8f09c4a2b65f"} Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.251623 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6ffd468-xgk27" event={"ID":"292d8e1e-de3b-4259-afef-c19e2c21187c","Type":"ContainerStarted","Data":"33432f8fb31556de3b070d1c39482bcc6d6666ff0bb7281f5a085f52c18b5f5d"} Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.252457 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.252508 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.253532 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hh5s2" podUID="426b37f9-4acb-4d28-a55c-3fcde2600002" containerName="registry-server" containerID="cri-o://3d2681901d856e71e99e2f5dd2205106a0fb9176c61fcdad6f85c78cea4118e6" gracePeriod=2 Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.272744 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-76d486c65f-w9lc6" podStartSLOduration=1.9453027710000002 podStartE2EDuration="5.272717884s" podCreationTimestamp="2026-02-20 08:23:06 +0000 UTC" firstStartedPulling="2026-02-20 08:23:07.095361807 +0000 UTC m=+1036.069856627" lastFinishedPulling="2026-02-20 08:23:10.42277693 +0000 UTC m=+1039.397271740" observedRunningTime="2026-02-20 08:23:11.247873779 +0000 UTC m=+1040.222368629" watchObservedRunningTime="2026-02-20 08:23:11.272717884 +0000 UTC m=+1040.247212714" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.309301 4948 scope.go:117] "RemoveContainer" containerID="61ecb15080679db6c7523b2b45faf0110c6e8bc826d5ac4d37af1077d531ed07" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.325924 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-6ffd468-xgk27" podStartSLOduration=2.061738272 podStartE2EDuration="5.32589663s" podCreationTimestamp="2026-02-20 08:23:06 +0000 UTC" firstStartedPulling="2026-02-20 08:23:07.169285536 +0000 UTC m=+1036.143780356" lastFinishedPulling="2026-02-20 08:23:10.433443894 +0000 UTC m=+1039.407938714" observedRunningTime="2026-02-20 08:23:11.279224405 +0000 UTC m=+1040.253719265" watchObservedRunningTime="2026-02-20 08:23:11.32589663 +0000 UTC m=+1040.300391450" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.352575 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.358845 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.412135 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.433454 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:11 crc kubenswrapper[4948]: E0220 08:23:11.433938 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89b49c50-bae7-47fe-b85f-7a718953933d" containerName="sg-core" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.433960 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="89b49c50-bae7-47fe-b85f-7a718953933d" containerName="sg-core" Feb 20 08:23:11 crc kubenswrapper[4948]: E0220 08:23:11.433992 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89b49c50-bae7-47fe-b85f-7a718953933d" containerName="ceilometer-notification-agent" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.434001 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="89b49c50-bae7-47fe-b85f-7a718953933d" containerName="ceilometer-notification-agent" Feb 20 08:23:11 crc kubenswrapper[4948]: E0220 08:23:11.434020 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e" containerName="init" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.434028 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e" containerName="init" Feb 20 08:23:11 crc kubenswrapper[4948]: E0220 08:23:11.434043 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89b49c50-bae7-47fe-b85f-7a718953933d" containerName="ceilometer-central-agent" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.434050 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="89b49c50-bae7-47fe-b85f-7a718953933d" containerName="ceilometer-central-agent" Feb 20 08:23:11 crc kubenswrapper[4948]: E0220 08:23:11.434069 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89b49c50-bae7-47fe-b85f-7a718953933d" containerName="proxy-httpd" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.434075 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="89b49c50-bae7-47fe-b85f-7a718953933d" containerName="proxy-httpd" Feb 20 08:23:11 crc kubenswrapper[4948]: E0220 08:23:11.434099 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ff6993b-2bde-429d-b0c7-cdea51f733a2" containerName="cinder-api" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.434107 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ff6993b-2bde-429d-b0c7-cdea51f733a2" containerName="cinder-api" Feb 20 08:23:11 crc kubenswrapper[4948]: E0220 08:23:11.434124 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ff6993b-2bde-429d-b0c7-cdea51f733a2" containerName="cinder-api-log" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.434131 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ff6993b-2bde-429d-b0c7-cdea51f733a2" containerName="cinder-api-log" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.434329 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="89b49c50-bae7-47fe-b85f-7a718953933d" containerName="ceilometer-central-agent" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.434343 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="89b49c50-bae7-47fe-b85f-7a718953933d" containerName="sg-core" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.434358 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e" containerName="init" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.434375 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="89b49c50-bae7-47fe-b85f-7a718953933d" containerName="proxy-httpd" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.434388 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ff6993b-2bde-429d-b0c7-cdea51f733a2" containerName="cinder-api-log" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.434402 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ff6993b-2bde-429d-b0c7-cdea51f733a2" containerName="cinder-api" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.434417 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="89b49c50-bae7-47fe-b85f-7a718953933d" containerName="ceilometer-notification-agent" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.436368 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.440350 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.443125 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.477142 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.505455 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.546057 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.547983 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.551823 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.552082 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.552301 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.557029 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.567242 4948 scope.go:117] "RemoveContainer" containerID="68df1bb1b13a5e62c745d68de5298b184c0cdcaa037a022eedeb800b70856493" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.580538 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-scripts\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.580597 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-config-data\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.580739 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.580954 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rkd9\" (UniqueName: \"kubernetes.io/projected/44092649-be82-4225-8cfc-d37dbe840877-kube-api-access-4rkd9\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.581203 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.581353 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/44092649-be82-4225-8cfc-d37dbe840877-run-httpd\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.581524 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/44092649-be82-4225-8cfc-d37dbe840877-log-httpd\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.622137 4948 scope.go:117] "RemoveContainer" containerID="f21ea8eddb7fd63986bb5357a809bc7c94ba2e30ea5976b9795bdbd7d103ba56" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.659076 4948 scope.go:117] "RemoveContainer" containerID="27fa5fd7b38b4d8b359bee17f753923040dddd64a704b2cd2a15217dcb4497e9" Feb 20 08:23:11 crc kubenswrapper[4948]: E0220 08:23:11.659577 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27fa5fd7b38b4d8b359bee17f753923040dddd64a704b2cd2a15217dcb4497e9\": container with ID starting with 27fa5fd7b38b4d8b359bee17f753923040dddd64a704b2cd2a15217dcb4497e9 not found: ID does not exist" containerID="27fa5fd7b38b4d8b359bee17f753923040dddd64a704b2cd2a15217dcb4497e9" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.659623 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27fa5fd7b38b4d8b359bee17f753923040dddd64a704b2cd2a15217dcb4497e9"} err="failed to get container status \"27fa5fd7b38b4d8b359bee17f753923040dddd64a704b2cd2a15217dcb4497e9\": rpc error: code = NotFound desc = could not find container \"27fa5fd7b38b4d8b359bee17f753923040dddd64a704b2cd2a15217dcb4497e9\": container with ID starting with 27fa5fd7b38b4d8b359bee17f753923040dddd64a704b2cd2a15217dcb4497e9 not found: ID does not exist" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.659651 4948 scope.go:117] "RemoveContainer" containerID="61ecb15080679db6c7523b2b45faf0110c6e8bc826d5ac4d37af1077d531ed07" Feb 20 08:23:11 crc kubenswrapper[4948]: E0220 08:23:11.663061 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61ecb15080679db6c7523b2b45faf0110c6e8bc826d5ac4d37af1077d531ed07\": container with ID starting with 61ecb15080679db6c7523b2b45faf0110c6e8bc826d5ac4d37af1077d531ed07 not found: ID does not exist" containerID="61ecb15080679db6c7523b2b45faf0110c6e8bc826d5ac4d37af1077d531ed07" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.663101 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61ecb15080679db6c7523b2b45faf0110c6e8bc826d5ac4d37af1077d531ed07"} err="failed to get container status \"61ecb15080679db6c7523b2b45faf0110c6e8bc826d5ac4d37af1077d531ed07\": rpc error: code = NotFound desc = could not find container \"61ecb15080679db6c7523b2b45faf0110c6e8bc826d5ac4d37af1077d531ed07\": container with ID starting with 61ecb15080679db6c7523b2b45faf0110c6e8bc826d5ac4d37af1077d531ed07 not found: ID does not exist" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.663127 4948 scope.go:117] "RemoveContainer" containerID="68df1bb1b13a5e62c745d68de5298b184c0cdcaa037a022eedeb800b70856493" Feb 20 08:23:11 crc kubenswrapper[4948]: E0220 08:23:11.663496 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68df1bb1b13a5e62c745d68de5298b184c0cdcaa037a022eedeb800b70856493\": container with ID starting with 68df1bb1b13a5e62c745d68de5298b184c0cdcaa037a022eedeb800b70856493 not found: ID does not exist" containerID="68df1bb1b13a5e62c745d68de5298b184c0cdcaa037a022eedeb800b70856493" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.663522 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68df1bb1b13a5e62c745d68de5298b184c0cdcaa037a022eedeb800b70856493"} err="failed to get container status \"68df1bb1b13a5e62c745d68de5298b184c0cdcaa037a022eedeb800b70856493\": rpc error: code = NotFound desc = could not find container \"68df1bb1b13a5e62c745d68de5298b184c0cdcaa037a022eedeb800b70856493\": container with ID starting with 68df1bb1b13a5e62c745d68de5298b184c0cdcaa037a022eedeb800b70856493 not found: ID does not exist" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.663538 4948 scope.go:117] "RemoveContainer" containerID="f21ea8eddb7fd63986bb5357a809bc7c94ba2e30ea5976b9795bdbd7d103ba56" Feb 20 08:23:11 crc kubenswrapper[4948]: E0220 08:23:11.668160 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f21ea8eddb7fd63986bb5357a809bc7c94ba2e30ea5976b9795bdbd7d103ba56\": container with ID starting with f21ea8eddb7fd63986bb5357a809bc7c94ba2e30ea5976b9795bdbd7d103ba56 not found: ID does not exist" containerID="f21ea8eddb7fd63986bb5357a809bc7c94ba2e30ea5976b9795bdbd7d103ba56" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.668200 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f21ea8eddb7fd63986bb5357a809bc7c94ba2e30ea5976b9795bdbd7d103ba56"} err="failed to get container status \"f21ea8eddb7fd63986bb5357a809bc7c94ba2e30ea5976b9795bdbd7d103ba56\": rpc error: code = NotFound desc = could not find container \"f21ea8eddb7fd63986bb5357a809bc7c94ba2e30ea5976b9795bdbd7d103ba56\": container with ID starting with f21ea8eddb7fd63986bb5357a809bc7c94ba2e30ea5976b9795bdbd7d103ba56 not found: ID does not exist" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.668225 4948 scope.go:117] "RemoveContainer" containerID="0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.683841 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/44092649-be82-4225-8cfc-d37dbe840877-run-httpd\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.683887 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-scripts\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.683928 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/44092649-be82-4225-8cfc-d37dbe840877-log-httpd\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.683956 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-config-data-custom\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.683988 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcn5v\" (UniqueName: \"kubernetes.io/projected/e5967cab-cc29-4c87-b873-0f5d20203d4f-kube-api-access-mcn5v\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.684005 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-scripts\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.684027 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-config-data\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.684046 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-config-data\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.684061 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.684091 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e5967cab-cc29-4c87-b873-0f5d20203d4f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.684118 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rkd9\" (UniqueName: \"kubernetes.io/projected/44092649-be82-4225-8cfc-d37dbe840877-kube-api-access-4rkd9\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.684145 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.684173 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5967cab-cc29-4c87-b873-0f5d20203d4f-logs\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.684192 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.684215 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.684235 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.684659 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/44092649-be82-4225-8cfc-d37dbe840877-run-httpd\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.684881 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/44092649-be82-4225-8cfc-d37dbe840877-log-httpd\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.685999 4948 scope.go:117] "RemoveContainer" containerID="2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.691602 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.691941 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-scripts\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.693492 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-config-data\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.708480 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.716770 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rkd9\" (UniqueName: \"kubernetes.io/projected/44092649-be82-4225-8cfc-d37dbe840877-kube-api-access-4rkd9\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.741322 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ff6993b-2bde-429d-b0c7-cdea51f733a2" path="/var/lib/kubelet/pods/1ff6993b-2bde-429d-b0c7-cdea51f733a2/volumes" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.742076 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89b49c50-bae7-47fe-b85f-7a718953933d" path="/var/lib/kubelet/pods/89b49c50-bae7-47fe-b85f-7a718953933d/volumes" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.781485 4948 scope.go:117] "RemoveContainer" containerID="0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3" Feb 20 08:23:11 crc kubenswrapper[4948]: E0220 08:23:11.781967 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3\": container with ID starting with 0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3 not found: ID does not exist" containerID="0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.782030 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3"} err="failed to get container status \"0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3\": rpc error: code = NotFound desc = could not find container \"0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3\": container with ID starting with 0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3 not found: ID does not exist" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.782072 4948 scope.go:117] "RemoveContainer" containerID="2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d" Feb 20 08:23:11 crc kubenswrapper[4948]: E0220 08:23:11.782556 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d\": container with ID starting with 2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d not found: ID does not exist" containerID="2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.782596 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d"} err="failed to get container status \"2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d\": rpc error: code = NotFound desc = could not find container \"2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d\": container with ID starting with 2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d not found: ID does not exist" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.782628 4948 scope.go:117] "RemoveContainer" containerID="0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.783153 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3"} err="failed to get container status \"0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3\": rpc error: code = NotFound desc = could not find container \"0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3\": container with ID starting with 0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3 not found: ID does not exist" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.783187 4948 scope.go:117] "RemoveContainer" containerID="2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.783485 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d"} err="failed to get container status \"2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d\": rpc error: code = NotFound desc = could not find container \"2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d\": container with ID starting with 2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d not found: ID does not exist" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.785682 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-config-data-custom\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.785752 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcn5v\" (UniqueName: \"kubernetes.io/projected/e5967cab-cc29-4c87-b873-0f5d20203d4f-kube-api-access-mcn5v\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.786320 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-config-data\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.787161 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e5967cab-cc29-4c87-b873-0f5d20203d4f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.787238 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e5967cab-cc29-4c87-b873-0f5d20203d4f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.787317 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.787850 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5967cab-cc29-4c87-b873-0f5d20203d4f-logs\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.787889 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.787922 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.788025 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-scripts\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.788325 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5967cab-cc29-4c87-b873-0f5d20203d4f-logs\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.789197 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-config-data-custom\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.790853 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-config-data\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.791870 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.792592 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.793187 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.794289 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-scripts\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.802225 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcn5v\" (UniqueName: \"kubernetes.io/projected/e5967cab-cc29-4c87-b873-0f5d20203d4f-kube-api-access-mcn5v\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.846563 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hh5s2" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.865965 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.889493 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.936224 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.992048 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/426b37f9-4acb-4d28-a55c-3fcde2600002-utilities\") pod \"426b37f9-4acb-4d28-a55c-3fcde2600002\" (UID: \"426b37f9-4acb-4d28-a55c-3fcde2600002\") " Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.992323 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/426b37f9-4acb-4d28-a55c-3fcde2600002-catalog-content\") pod \"426b37f9-4acb-4d28-a55c-3fcde2600002\" (UID: \"426b37f9-4acb-4d28-a55c-3fcde2600002\") " Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.992428 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tw6bb\" (UniqueName: \"kubernetes.io/projected/426b37f9-4acb-4d28-a55c-3fcde2600002-kube-api-access-tw6bb\") pod \"426b37f9-4acb-4d28-a55c-3fcde2600002\" (UID: \"426b37f9-4acb-4d28-a55c-3fcde2600002\") " Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.992793 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/426b37f9-4acb-4d28-a55c-3fcde2600002-utilities" (OuterVolumeSpecName: "utilities") pod "426b37f9-4acb-4d28-a55c-3fcde2600002" (UID: "426b37f9-4acb-4d28-a55c-3fcde2600002"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.997281 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/426b37f9-4acb-4d28-a55c-3fcde2600002-kube-api-access-tw6bb" (OuterVolumeSpecName: "kube-api-access-tw6bb") pod "426b37f9-4acb-4d28-a55c-3fcde2600002" (UID: "426b37f9-4acb-4d28-a55c-3fcde2600002"). InnerVolumeSpecName "kube-api-access-tw6bb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.079572 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/426b37f9-4acb-4d28-a55c-3fcde2600002-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "426b37f9-4acb-4d28-a55c-3fcde2600002" (UID: "426b37f9-4acb-4d28-a55c-3fcde2600002"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.096477 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/426b37f9-4acb-4d28-a55c-3fcde2600002-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.096538 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tw6bb\" (UniqueName: \"kubernetes.io/projected/426b37f9-4acb-4d28-a55c-3fcde2600002-kube-api-access-tw6bb\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.096550 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/426b37f9-4acb-4d28-a55c-3fcde2600002-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.136798 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-686d666564-8tnjb"] Feb 20 08:23:12 crc kubenswrapper[4948]: E0220 08:23:12.137395 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="426b37f9-4acb-4d28-a55c-3fcde2600002" containerName="registry-server" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.137410 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="426b37f9-4acb-4d28-a55c-3fcde2600002" containerName="registry-server" Feb 20 08:23:12 crc kubenswrapper[4948]: E0220 08:23:12.137421 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="426b37f9-4acb-4d28-a55c-3fcde2600002" containerName="extract-utilities" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.137427 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="426b37f9-4acb-4d28-a55c-3fcde2600002" containerName="extract-utilities" Feb 20 08:23:12 crc kubenswrapper[4948]: E0220 08:23:12.137447 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="426b37f9-4acb-4d28-a55c-3fcde2600002" containerName="extract-content" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.137453 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="426b37f9-4acb-4d28-a55c-3fcde2600002" containerName="extract-content" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.137607 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="426b37f9-4acb-4d28-a55c-3fcde2600002" containerName="registry-server" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.138479 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.141797 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.143101 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.170835 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-686d666564-8tnjb"] Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.199026 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/51f6301c-57be-4a07-91df-b1a9f90e3f40-internal-tls-certs\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.199103 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51f6301c-57be-4a07-91df-b1a9f90e3f40-config-data-custom\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.199156 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/51f6301c-57be-4a07-91df-b1a9f90e3f40-public-tls-certs\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.199205 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51f6301c-57be-4a07-91df-b1a9f90e3f40-config-data\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.199227 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51f6301c-57be-4a07-91df-b1a9f90e3f40-logs\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.199255 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51f6301c-57be-4a07-91df-b1a9f90e3f40-combined-ca-bundle\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.199274 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fdqm\" (UniqueName: \"kubernetes.io/projected/51f6301c-57be-4a07-91df-b1a9f90e3f40-kube-api-access-5fdqm\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.292240 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-685bb7f5c-jd9nm"] Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.292465 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-685bb7f5c-jd9nm" podUID="e0661780-ca2b-42ae-be25-5d0e914fbb49" containerName="neutron-api" containerID="cri-o://1ad25f5bfef4b0fa06a939cc9fb8e5bfa3dad488d3265a3cd5fa655c097dda5b" gracePeriod=30 Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.292819 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-685bb7f5c-jd9nm" podUID="e0661780-ca2b-42ae-be25-5d0e914fbb49" containerName="neutron-httpd" containerID="cri-o://d2ca1451e75be7daafc6907a8305baa229ed6441f1f79895bd2df00e5c17bfa7" gracePeriod=30 Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.300838 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/51f6301c-57be-4a07-91df-b1a9f90e3f40-public-tls-certs\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.300937 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51f6301c-57be-4a07-91df-b1a9f90e3f40-config-data\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.300982 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51f6301c-57be-4a07-91df-b1a9f90e3f40-logs\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.301030 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51f6301c-57be-4a07-91df-b1a9f90e3f40-combined-ca-bundle\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.301056 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fdqm\" (UniqueName: \"kubernetes.io/projected/51f6301c-57be-4a07-91df-b1a9f90e3f40-kube-api-access-5fdqm\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.301110 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/51f6301c-57be-4a07-91df-b1a9f90e3f40-internal-tls-certs\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.301167 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51f6301c-57be-4a07-91df-b1a9f90e3f40-config-data-custom\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.301574 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51f6301c-57be-4a07-91df-b1a9f90e3f40-logs\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.305303 4948 generic.go:334] "Generic (PLEG): container finished" podID="426b37f9-4acb-4d28-a55c-3fcde2600002" containerID="3d2681901d856e71e99e2f5dd2205106a0fb9176c61fcdad6f85c78cea4118e6" exitCode=0 Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.308367 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hh5s2" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.309230 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hh5s2" event={"ID":"426b37f9-4acb-4d28-a55c-3fcde2600002","Type":"ContainerDied","Data":"3d2681901d856e71e99e2f5dd2205106a0fb9176c61fcdad6f85c78cea4118e6"} Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.309268 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hh5s2" event={"ID":"426b37f9-4acb-4d28-a55c-3fcde2600002","Type":"ContainerDied","Data":"6c8a403f4e4cb34b74e5da72f8bdf18b9a2a922052dce8500240a25dba684641"} Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.309286 4948 scope.go:117] "RemoveContainer" containerID="3d2681901d856e71e99e2f5dd2205106a0fb9176c61fcdad6f85c78cea4118e6" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.317187 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51f6301c-57be-4a07-91df-b1a9f90e3f40-config-data\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.319653 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/51f6301c-57be-4a07-91df-b1a9f90e3f40-public-tls-certs\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.323265 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-685bb7f5c-jd9nm" podUID="e0661780-ca2b-42ae-be25-5d0e914fbb49" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.156:9696/\": EOF" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.326784 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51f6301c-57be-4a07-91df-b1a9f90e3f40-config-data-custom\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.336929 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fdqm\" (UniqueName: \"kubernetes.io/projected/51f6301c-57be-4a07-91df-b1a9f90e3f40-kube-api-access-5fdqm\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.342495 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51f6301c-57be-4a07-91df-b1a9f90e3f40-combined-ca-bundle\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.358918 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-76b5684765-h5624"] Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.360330 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.364297 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/51f6301c-57be-4a07-91df-b1a9f90e3f40-internal-tls-certs\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.379083 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-76b5684765-h5624"] Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.443895 4948 scope.go:117] "RemoveContainer" containerID="2d58a1beb2801a36e08660bfb99639b7abb52f4d18887bdac3f68efb75c8316d" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.444021 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.454767 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hh5s2"] Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.462664 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hh5s2"] Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.465925 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: W0220 08:23:12.504113 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5967cab_cc29_4c87_b873_0f5d20203d4f.slice/crio-29185c36a66efe969c66c1ca5274cb67c9aad81d503060845b8062dd34d95951 WatchSource:0}: Error finding container 29185c36a66efe969c66c1ca5274cb67c9aad81d503060845b8062dd34d95951: Status 404 returned error can't find the container with id 29185c36a66efe969c66c1ca5274cb67c9aad81d503060845b8062dd34d95951 Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.510171 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-internal-tls-certs\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.510359 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-ovndb-tls-certs\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.510575 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-httpd-config\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.510698 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-combined-ca-bundle\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.510933 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-public-tls-certs\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.511085 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f67zh\" (UniqueName: \"kubernetes.io/projected/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-kube-api-access-f67zh\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.511265 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-config\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.512361 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.582854 4948 scope.go:117] "RemoveContainer" containerID="62e9964b1041b8a3ee083b7749dd1ff49cbef0514cc265f8b36923ced11f9863" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.613017 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-config\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.613099 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-internal-tls-certs\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.613119 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-ovndb-tls-certs\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.613215 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-httpd-config\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.613258 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-combined-ca-bundle\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.613275 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-public-tls-certs\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.613328 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f67zh\" (UniqueName: \"kubernetes.io/projected/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-kube-api-access-f67zh\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.618301 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-ovndb-tls-certs\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.618390 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-config\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.622753 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-combined-ca-bundle\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.627635 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-public-tls-certs\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.629818 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-httpd-config\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.630508 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-internal-tls-certs\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.638591 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f67zh\" (UniqueName: \"kubernetes.io/projected/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-kube-api-access-f67zh\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.639104 4948 scope.go:117] "RemoveContainer" containerID="3d2681901d856e71e99e2f5dd2205106a0fb9176c61fcdad6f85c78cea4118e6" Feb 20 08:23:12 crc kubenswrapper[4948]: E0220 08:23:12.639493 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d2681901d856e71e99e2f5dd2205106a0fb9176c61fcdad6f85c78cea4118e6\": container with ID starting with 3d2681901d856e71e99e2f5dd2205106a0fb9176c61fcdad6f85c78cea4118e6 not found: ID does not exist" containerID="3d2681901d856e71e99e2f5dd2205106a0fb9176c61fcdad6f85c78cea4118e6" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.639575 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d2681901d856e71e99e2f5dd2205106a0fb9176c61fcdad6f85c78cea4118e6"} err="failed to get container status \"3d2681901d856e71e99e2f5dd2205106a0fb9176c61fcdad6f85c78cea4118e6\": rpc error: code = NotFound desc = could not find container \"3d2681901d856e71e99e2f5dd2205106a0fb9176c61fcdad6f85c78cea4118e6\": container with ID starting with 3d2681901d856e71e99e2f5dd2205106a0fb9176c61fcdad6f85c78cea4118e6 not found: ID does not exist" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.639644 4948 scope.go:117] "RemoveContainer" containerID="2d58a1beb2801a36e08660bfb99639b7abb52f4d18887bdac3f68efb75c8316d" Feb 20 08:23:12 crc kubenswrapper[4948]: E0220 08:23:12.639898 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d58a1beb2801a36e08660bfb99639b7abb52f4d18887bdac3f68efb75c8316d\": container with ID starting with 2d58a1beb2801a36e08660bfb99639b7abb52f4d18887bdac3f68efb75c8316d not found: ID does not exist" containerID="2d58a1beb2801a36e08660bfb99639b7abb52f4d18887bdac3f68efb75c8316d" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.639982 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d58a1beb2801a36e08660bfb99639b7abb52f4d18887bdac3f68efb75c8316d"} err="failed to get container status \"2d58a1beb2801a36e08660bfb99639b7abb52f4d18887bdac3f68efb75c8316d\": rpc error: code = NotFound desc = could not find container \"2d58a1beb2801a36e08660bfb99639b7abb52f4d18887bdac3f68efb75c8316d\": container with ID starting with 2d58a1beb2801a36e08660bfb99639b7abb52f4d18887bdac3f68efb75c8316d not found: ID does not exist" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.640053 4948 scope.go:117] "RemoveContainer" containerID="62e9964b1041b8a3ee083b7749dd1ff49cbef0514cc265f8b36923ced11f9863" Feb 20 08:23:12 crc kubenswrapper[4948]: E0220 08:23:12.644587 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62e9964b1041b8a3ee083b7749dd1ff49cbef0514cc265f8b36923ced11f9863\": container with ID starting with 62e9964b1041b8a3ee083b7749dd1ff49cbef0514cc265f8b36923ced11f9863 not found: ID does not exist" containerID="62e9964b1041b8a3ee083b7749dd1ff49cbef0514cc265f8b36923ced11f9863" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.644635 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62e9964b1041b8a3ee083b7749dd1ff49cbef0514cc265f8b36923ced11f9863"} err="failed to get container status \"62e9964b1041b8a3ee083b7749dd1ff49cbef0514cc265f8b36923ced11f9863\": rpc error: code = NotFound desc = could not find container \"62e9964b1041b8a3ee083b7749dd1ff49cbef0514cc265f8b36923ced11f9863\": container with ID starting with 62e9964b1041b8a3ee083b7749dd1ff49cbef0514cc265f8b36923ced11f9863 not found: ID does not exist" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.740303 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.940486 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-686d666564-8tnjb"] Feb 20 08:23:12 crc kubenswrapper[4948]: W0220 08:23:12.966317 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51f6301c_57be_4a07_91df_b1a9f90e3f40.slice/crio-af0403aeb9d6e04d55a315e4c351060deae309bea1ede2db1e668df78c7110a7 WatchSource:0}: Error finding container af0403aeb9d6e04d55a315e4c351060deae309bea1ede2db1e668df78c7110a7: Status 404 returned error can't find the container with id af0403aeb9d6e04d55a315e4c351060deae309bea1ede2db1e668df78c7110a7 Feb 20 08:23:13 crc kubenswrapper[4948]: I0220 08:23:13.282799 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-76b5684765-h5624"] Feb 20 08:23:13 crc kubenswrapper[4948]: I0220 08:23:13.321411 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-686d666564-8tnjb" event={"ID":"51f6301c-57be-4a07-91df-b1a9f90e3f40","Type":"ContainerStarted","Data":"8d925a9cf348b2c3c2fa4ee34cb7d0ae0a9d07d119fd8e257a3a9e7f5a687bbd"} Feb 20 08:23:13 crc kubenswrapper[4948]: I0220 08:23:13.321453 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-686d666564-8tnjb" event={"ID":"51f6301c-57be-4a07-91df-b1a9f90e3f40","Type":"ContainerStarted","Data":"af0403aeb9d6e04d55a315e4c351060deae309bea1ede2db1e668df78c7110a7"} Feb 20 08:23:13 crc kubenswrapper[4948]: I0220 08:23:13.325234 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-76b5684765-h5624" event={"ID":"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59","Type":"ContainerStarted","Data":"38aa7807e10bebc2e535552cb2dda45a3e1972a20649904b4402700bc8d8dd22"} Feb 20 08:23:13 crc kubenswrapper[4948]: I0220 08:23:13.328032 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"44092649-be82-4225-8cfc-d37dbe840877","Type":"ContainerStarted","Data":"eab93963947d8aecd2b45387be911464a6576e8236c9ccd3d0a009ad65054a74"} Feb 20 08:23:13 crc kubenswrapper[4948]: I0220 08:23:13.328150 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"44092649-be82-4225-8cfc-d37dbe840877","Type":"ContainerStarted","Data":"8266937b599da958a185777fc1969453e86491578f18e8d8f35aa2e0681bba26"} Feb 20 08:23:13 crc kubenswrapper[4948]: I0220 08:23:13.329317 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e5967cab-cc29-4c87-b873-0f5d20203d4f","Type":"ContainerStarted","Data":"af195efb8dccb5b78426393f4467711f39f97b1e3850500984806b51282213f6"} Feb 20 08:23:13 crc kubenswrapper[4948]: I0220 08:23:13.329451 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e5967cab-cc29-4c87-b873-0f5d20203d4f","Type":"ContainerStarted","Data":"29185c36a66efe969c66c1ca5274cb67c9aad81d503060845b8062dd34d95951"} Feb 20 08:23:13 crc kubenswrapper[4948]: I0220 08:23:13.370949 4948 generic.go:334] "Generic (PLEG): container finished" podID="e0661780-ca2b-42ae-be25-5d0e914fbb49" containerID="d2ca1451e75be7daafc6907a8305baa229ed6441f1f79895bd2df00e5c17bfa7" exitCode=0 Feb 20 08:23:13 crc kubenswrapper[4948]: I0220 08:23:13.371178 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-685bb7f5c-jd9nm" event={"ID":"e0661780-ca2b-42ae-be25-5d0e914fbb49","Type":"ContainerDied","Data":"d2ca1451e75be7daafc6907a8305baa229ed6441f1f79895bd2df00e5c17bfa7"} Feb 20 08:23:13 crc kubenswrapper[4948]: I0220 08:23:13.737440 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="426b37f9-4acb-4d28-a55c-3fcde2600002" path="/var/lib/kubelet/pods/426b37f9-4acb-4d28-a55c-3fcde2600002/volumes" Feb 20 08:23:13 crc kubenswrapper[4948]: I0220 08:23:13.927583 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-685bb7f5c-jd9nm" podUID="e0661780-ca2b-42ae-be25-5d0e914fbb49" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.156:9696/\": dial tcp 10.217.0.156:9696: connect: connection refused" Feb 20 08:23:14 crc kubenswrapper[4948]: I0220 08:23:14.067568 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:14 crc kubenswrapper[4948]: I0220 08:23:14.379005 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e5967cab-cc29-4c87-b873-0f5d20203d4f","Type":"ContainerStarted","Data":"e49db6d27fc81d3422ce0d425640ec2f5701bf916f725129edc6e1101abe2979"} Feb 20 08:23:14 crc kubenswrapper[4948]: I0220 08:23:14.380015 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 20 08:23:14 crc kubenswrapper[4948]: I0220 08:23:14.382237 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-686d666564-8tnjb" event={"ID":"51f6301c-57be-4a07-91df-b1a9f90e3f40","Type":"ContainerStarted","Data":"9d42d5115f0ed851367331fc56d5ac48b9ae3f6d9455032d65b1f2b001c75c79"} Feb 20 08:23:14 crc kubenswrapper[4948]: I0220 08:23:14.382600 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:14 crc kubenswrapper[4948]: I0220 08:23:14.384265 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-76b5684765-h5624" event={"ID":"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59","Type":"ContainerStarted","Data":"09aa54f86c858f9fc51ae59c1879f3a7cbc7a7847eacf1ae0b7ae4f33b0273fc"} Feb 20 08:23:14 crc kubenswrapper[4948]: I0220 08:23:14.384311 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-76b5684765-h5624" event={"ID":"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59","Type":"ContainerStarted","Data":"2ae99d6a857d168ec7f201c1bbf30b4bd88e112bcf930306d6a6abfbab050ad0"} Feb 20 08:23:14 crc kubenswrapper[4948]: I0220 08:23:14.384466 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:14 crc kubenswrapper[4948]: I0220 08:23:14.386464 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"44092649-be82-4225-8cfc-d37dbe840877","Type":"ContainerStarted","Data":"7e6d02d6234c2e72678f463fac466a22ef0548ef1c161d0d356079f01463ff09"} Feb 20 08:23:14 crc kubenswrapper[4948]: I0220 08:23:14.413582 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.41355432 podStartE2EDuration="3.41355432s" podCreationTimestamp="2026-02-20 08:23:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:23:14.406642149 +0000 UTC m=+1043.381136969" watchObservedRunningTime="2026-02-20 08:23:14.41355432 +0000 UTC m=+1043.388049140" Feb 20 08:23:14 crc kubenswrapper[4948]: I0220 08:23:14.437961 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-686d666564-8tnjb" podStartSLOduration=2.437940353 podStartE2EDuration="2.437940353s" podCreationTimestamp="2026-02-20 08:23:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:23:14.430391997 +0000 UTC m=+1043.404886817" watchObservedRunningTime="2026-02-20 08:23:14.437940353 +0000 UTC m=+1043.412435173" Feb 20 08:23:14 crc kubenswrapper[4948]: I0220 08:23:14.461361 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-76b5684765-h5624" podStartSLOduration=2.461341902 podStartE2EDuration="2.461341902s" podCreationTimestamp="2026-02-20 08:23:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:23:14.45639865 +0000 UTC m=+1043.430893470" watchObservedRunningTime="2026-02-20 08:23:14.461341902 +0000 UTC m=+1043.435836722" Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.396959 4948 generic.go:334] "Generic (PLEG): container finished" podID="e0661780-ca2b-42ae-be25-5d0e914fbb49" containerID="1ad25f5bfef4b0fa06a939cc9fb8e5bfa3dad488d3265a3cd5fa655c097dda5b" exitCode=0 Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.397050 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-685bb7f5c-jd9nm" event={"ID":"e0661780-ca2b-42ae-be25-5d0e914fbb49","Type":"ContainerDied","Data":"1ad25f5bfef4b0fa06a939cc9fb8e5bfa3dad488d3265a3cd5fa655c097dda5b"} Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.401482 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"44092649-be82-4225-8cfc-d37dbe840877","Type":"ContainerStarted","Data":"d829cc037c7ba0bd0abfe863863495a3388592344985fae656879df8c40ec990"} Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.401912 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.555428 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.766702 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.811740 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.846679 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.876568 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-config\") pod \"e0661780-ca2b-42ae-be25-5d0e914fbb49\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.876629 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-internal-tls-certs\") pod \"e0661780-ca2b-42ae-be25-5d0e914fbb49\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.876677 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-combined-ca-bundle\") pod \"e0661780-ca2b-42ae-be25-5d0e914fbb49\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.876732 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-public-tls-certs\") pod \"e0661780-ca2b-42ae-be25-5d0e914fbb49\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.876779 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-ovndb-tls-certs\") pod \"e0661780-ca2b-42ae-be25-5d0e914fbb49\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.876840 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fh687\" (UniqueName: \"kubernetes.io/projected/e0661780-ca2b-42ae-be25-5d0e914fbb49-kube-api-access-fh687\") pod \"e0661780-ca2b-42ae-be25-5d0e914fbb49\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.876868 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-httpd-config\") pod \"e0661780-ca2b-42ae-be25-5d0e914fbb49\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.884075 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0661780-ca2b-42ae-be25-5d0e914fbb49-kube-api-access-fh687" (OuterVolumeSpecName: "kube-api-access-fh687") pod "e0661780-ca2b-42ae-be25-5d0e914fbb49" (UID: "e0661780-ca2b-42ae-be25-5d0e914fbb49"). InnerVolumeSpecName "kube-api-access-fh687". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.929115 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "e0661780-ca2b-42ae-be25-5d0e914fbb49" (UID: "e0661780-ca2b-42ae-be25-5d0e914fbb49"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.942418 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e0661780-ca2b-42ae-be25-5d0e914fbb49" (UID: "e0661780-ca2b-42ae-be25-5d0e914fbb49"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.966054 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e0661780-ca2b-42ae-be25-5d0e914fbb49" (UID: "e0661780-ca2b-42ae-be25-5d0e914fbb49"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.978463 4948 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.978491 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.978501 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fh687\" (UniqueName: \"kubernetes.io/projected/e0661780-ca2b-42ae-be25-5d0e914fbb49-kube-api-access-fh687\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.978512 4948 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.991235 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-config" (OuterVolumeSpecName: "config") pod "e0661780-ca2b-42ae-be25-5d0e914fbb49" (UID: "e0661780-ca2b-42ae-be25-5d0e914fbb49"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.994937 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "e0661780-ca2b-42ae-be25-5d0e914fbb49" (UID: "e0661780-ca2b-42ae-be25-5d0e914fbb49"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.005122 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e0661780-ca2b-42ae-be25-5d0e914fbb49" (UID: "e0661780-ca2b-42ae-be25-5d0e914fbb49"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.079863 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.080095 4948 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.080183 4948 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.416798 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"44092649-be82-4225-8cfc-d37dbe840877","Type":"ContainerStarted","Data":"e85736cf9a602b496a8372a3fe185272b09bf09e11882adec2e63f62c62bf3d9"} Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.416961 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.421098 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-685bb7f5c-jd9nm" event={"ID":"e0661780-ca2b-42ae-be25-5d0e914fbb49","Type":"ContainerDied","Data":"4885fc2e4d8fe9fc0c29512f6a5fd4f1b7a87efbf9c10fe17b8bf4fe7f1ca02a"} Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.421173 4948 scope.go:117] "RemoveContainer" containerID="d2ca1451e75be7daafc6907a8305baa229ed6441f1f79895bd2df00e5c17bfa7" Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.421235 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="75294aee-c7f3-4b0b-82fc-df058490b90f" containerName="cinder-scheduler" containerID="cri-o://0c8a24ef5e8c2ed45ddfc6908331dbda28dd167039cb8bc7f442c18afef4d964" gracePeriod=30 Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.421259 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.421272 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="75294aee-c7f3-4b0b-82fc-df058490b90f" containerName="probe" containerID="cri-o://c1fbafc4d60f4668a3ce40442dbec31bba82bcb8754ebb88a8aed726953af8e7" gracePeriod=30 Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.452945 4948 scope.go:117] "RemoveContainer" containerID="1ad25f5bfef4b0fa06a939cc9fb8e5bfa3dad488d3265a3cd5fa655c097dda5b" Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.457822 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.714538234 podStartE2EDuration="5.457802438s" podCreationTimestamp="2026-02-20 08:23:11 +0000 UTC" firstStartedPulling="2026-02-20 08:23:12.465250595 +0000 UTC m=+1041.439745415" lastFinishedPulling="2026-02-20 08:23:16.208514799 +0000 UTC m=+1045.183009619" observedRunningTime="2026-02-20 08:23:16.450094608 +0000 UTC m=+1045.424589438" watchObservedRunningTime="2026-02-20 08:23:16.457802438 +0000 UTC m=+1045.432297268" Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.473618 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-685bb7f5c-jd9nm"] Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.481095 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-685bb7f5c-jd9nm"] Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.693753 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.769187 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-v8n5p"] Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.770127 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" podUID="0cc4ba01-323f-4bcb-9bc3-d639514f6088" containerName="dnsmasq-dns" containerID="cri-o://ce123808c7ffa9e9792ff9cd3fa77559a772ebe4ee4f08ad46bdfe3f21ea2386" gracePeriod=10 Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.811550 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" podUID="0cc4ba01-323f-4bcb-9bc3-d639514f6088" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.154:5353: connect: connection refused" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.418926 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.433318 4948 generic.go:334] "Generic (PLEG): container finished" podID="75294aee-c7f3-4b0b-82fc-df058490b90f" containerID="c1fbafc4d60f4668a3ce40442dbec31bba82bcb8754ebb88a8aed726953af8e7" exitCode=0 Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.433393 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"75294aee-c7f3-4b0b-82fc-df058490b90f","Type":"ContainerDied","Data":"c1fbafc4d60f4668a3ce40442dbec31bba82bcb8754ebb88a8aed726953af8e7"} Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.439874 4948 generic.go:334] "Generic (PLEG): container finished" podID="0cc4ba01-323f-4bcb-9bc3-d639514f6088" containerID="ce123808c7ffa9e9792ff9cd3fa77559a772ebe4ee4f08ad46bdfe3f21ea2386" exitCode=0 Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.439993 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.439996 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" event={"ID":"0cc4ba01-323f-4bcb-9bc3-d639514f6088","Type":"ContainerDied","Data":"ce123808c7ffa9e9792ff9cd3fa77559a772ebe4ee4f08ad46bdfe3f21ea2386"} Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.440152 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" event={"ID":"0cc4ba01-323f-4bcb-9bc3-d639514f6088","Type":"ContainerDied","Data":"5ba21a5e281f6be663a46ae25fc6e9402ee9bcd0e6caec5c0c3b4658222b76c1"} Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.440188 4948 scope.go:117] "RemoveContainer" containerID="ce123808c7ffa9e9792ff9cd3fa77559a772ebe4ee4f08ad46bdfe3f21ea2386" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.470891 4948 scope.go:117] "RemoveContainer" containerID="7121b5f51a9b7787780e4d13a76cf2b72d32878e2e19ab652f2bed18d9c68743" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.502324 4948 scope.go:117] "RemoveContainer" containerID="ce123808c7ffa9e9792ff9cd3fa77559a772ebe4ee4f08ad46bdfe3f21ea2386" Feb 20 08:23:17 crc kubenswrapper[4948]: E0220 08:23:17.502715 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce123808c7ffa9e9792ff9cd3fa77559a772ebe4ee4f08ad46bdfe3f21ea2386\": container with ID starting with ce123808c7ffa9e9792ff9cd3fa77559a772ebe4ee4f08ad46bdfe3f21ea2386 not found: ID does not exist" containerID="ce123808c7ffa9e9792ff9cd3fa77559a772ebe4ee4f08ad46bdfe3f21ea2386" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.502753 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce123808c7ffa9e9792ff9cd3fa77559a772ebe4ee4f08ad46bdfe3f21ea2386"} err="failed to get container status \"ce123808c7ffa9e9792ff9cd3fa77559a772ebe4ee4f08ad46bdfe3f21ea2386\": rpc error: code = NotFound desc = could not find container \"ce123808c7ffa9e9792ff9cd3fa77559a772ebe4ee4f08ad46bdfe3f21ea2386\": container with ID starting with ce123808c7ffa9e9792ff9cd3fa77559a772ebe4ee4f08ad46bdfe3f21ea2386 not found: ID does not exist" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.502776 4948 scope.go:117] "RemoveContainer" containerID="7121b5f51a9b7787780e4d13a76cf2b72d32878e2e19ab652f2bed18d9c68743" Feb 20 08:23:17 crc kubenswrapper[4948]: E0220 08:23:17.503199 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7121b5f51a9b7787780e4d13a76cf2b72d32878e2e19ab652f2bed18d9c68743\": container with ID starting with 7121b5f51a9b7787780e4d13a76cf2b72d32878e2e19ab652f2bed18d9c68743 not found: ID does not exist" containerID="7121b5f51a9b7787780e4d13a76cf2b72d32878e2e19ab652f2bed18d9c68743" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.503235 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7121b5f51a9b7787780e4d13a76cf2b72d32878e2e19ab652f2bed18d9c68743"} err="failed to get container status \"7121b5f51a9b7787780e4d13a76cf2b72d32878e2e19ab652f2bed18d9c68743\": rpc error: code = NotFound desc = could not find container \"7121b5f51a9b7787780e4d13a76cf2b72d32878e2e19ab652f2bed18d9c68743\": container with ID starting with 7121b5f51a9b7787780e4d13a76cf2b72d32878e2e19ab652f2bed18d9c68743 not found: ID does not exist" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.509150 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-dns-swift-storage-0\") pod \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.509191 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgbxd\" (UniqueName: \"kubernetes.io/projected/0cc4ba01-323f-4bcb-9bc3-d639514f6088-kube-api-access-sgbxd\") pod \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.509246 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-dns-svc\") pod \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.509276 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-config\") pod \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.509488 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-ovsdbserver-sb\") pod \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.509515 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-ovsdbserver-nb\") pod \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.516297 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cc4ba01-323f-4bcb-9bc3-d639514f6088-kube-api-access-sgbxd" (OuterVolumeSpecName: "kube-api-access-sgbxd") pod "0cc4ba01-323f-4bcb-9bc3-d639514f6088" (UID: "0cc4ba01-323f-4bcb-9bc3-d639514f6088"). InnerVolumeSpecName "kube-api-access-sgbxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.554408 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0cc4ba01-323f-4bcb-9bc3-d639514f6088" (UID: "0cc4ba01-323f-4bcb-9bc3-d639514f6088"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.559311 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0cc4ba01-323f-4bcb-9bc3-d639514f6088" (UID: "0cc4ba01-323f-4bcb-9bc3-d639514f6088"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.560495 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0cc4ba01-323f-4bcb-9bc3-d639514f6088" (UID: "0cc4ba01-323f-4bcb-9bc3-d639514f6088"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.564240 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0cc4ba01-323f-4bcb-9bc3-d639514f6088" (UID: "0cc4ba01-323f-4bcb-9bc3-d639514f6088"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.577883 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-config" (OuterVolumeSpecName: "config") pod "0cc4ba01-323f-4bcb-9bc3-d639514f6088" (UID: "0cc4ba01-323f-4bcb-9bc3-d639514f6088"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.610956 4948 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.611011 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgbxd\" (UniqueName: \"kubernetes.io/projected/0cc4ba01-323f-4bcb-9bc3-d639514f6088-kube-api-access-sgbxd\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.611025 4948 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.611036 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.611046 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.611055 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.732528 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0661780-ca2b-42ae-be25-5d0e914fbb49" path="/var/lib/kubelet/pods/e0661780-ca2b-42ae-be25-5d0e914fbb49/volumes" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.774934 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-v8n5p"] Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.782470 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-v8n5p"] Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.454230 4948 generic.go:334] "Generic (PLEG): container finished" podID="75294aee-c7f3-4b0b-82fc-df058490b90f" containerID="0c8a24ef5e8c2ed45ddfc6908331dbda28dd167039cb8bc7f442c18afef4d964" exitCode=0 Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.454310 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"75294aee-c7f3-4b0b-82fc-df058490b90f","Type":"ContainerDied","Data":"0c8a24ef5e8c2ed45ddfc6908331dbda28dd167039cb8bc7f442c18afef4d964"} Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.828263 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.838607 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vllr\" (UniqueName: \"kubernetes.io/projected/75294aee-c7f3-4b0b-82fc-df058490b90f-kube-api-access-8vllr\") pod \"75294aee-c7f3-4b0b-82fc-df058490b90f\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.838672 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-config-data-custom\") pod \"75294aee-c7f3-4b0b-82fc-df058490b90f\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.838705 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-combined-ca-bundle\") pod \"75294aee-c7f3-4b0b-82fc-df058490b90f\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.838724 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-config-data\") pod \"75294aee-c7f3-4b0b-82fc-df058490b90f\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.838762 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-scripts\") pod \"75294aee-c7f3-4b0b-82fc-df058490b90f\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.838780 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/75294aee-c7f3-4b0b-82fc-df058490b90f-etc-machine-id\") pod \"75294aee-c7f3-4b0b-82fc-df058490b90f\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.839076 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/75294aee-c7f3-4b0b-82fc-df058490b90f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "75294aee-c7f3-4b0b-82fc-df058490b90f" (UID: "75294aee-c7f3-4b0b-82fc-df058490b90f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.847426 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75294aee-c7f3-4b0b-82fc-df058490b90f-kube-api-access-8vllr" (OuterVolumeSpecName: "kube-api-access-8vllr") pod "75294aee-c7f3-4b0b-82fc-df058490b90f" (UID: "75294aee-c7f3-4b0b-82fc-df058490b90f"). InnerVolumeSpecName "kube-api-access-8vllr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.907153 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "75294aee-c7f3-4b0b-82fc-df058490b90f" (UID: "75294aee-c7f3-4b0b-82fc-df058490b90f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.911186 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-scripts" (OuterVolumeSpecName: "scripts") pod "75294aee-c7f3-4b0b-82fc-df058490b90f" (UID: "75294aee-c7f3-4b0b-82fc-df058490b90f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.956922 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vllr\" (UniqueName: \"kubernetes.io/projected/75294aee-c7f3-4b0b-82fc-df058490b90f-kube-api-access-8vllr\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.956959 4948 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.957087 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.957100 4948 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/75294aee-c7f3-4b0b-82fc-df058490b90f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.988201 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "75294aee-c7f3-4b0b-82fc-df058490b90f" (UID: "75294aee-c7f3-4b0b-82fc-df058490b90f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.029410 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.062690 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.072581 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-config-data" (OuterVolumeSpecName: "config-data") pod "75294aee-c7f3-4b0b-82fc-df058490b90f" (UID: "75294aee-c7f3-4b0b-82fc-df058490b90f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.164645 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.313526 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.470556 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"75294aee-c7f3-4b0b-82fc-df058490b90f","Type":"ContainerDied","Data":"3927e56027e8aee8bf743a6fd812a761cbedf51a0e339d17550adee912677f5f"} Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.470623 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.470988 4948 scope.go:117] "RemoveContainer" containerID="c1fbafc4d60f4668a3ce40442dbec31bba82bcb8754ebb88a8aed726953af8e7" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.514465 4948 scope.go:117] "RemoveContainer" containerID="0c8a24ef5e8c2ed45ddfc6908331dbda28dd167039cb8bc7f442c18afef4d964" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.519846 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.530098 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.545873 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 20 08:23:19 crc kubenswrapper[4948]: E0220 08:23:19.546448 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cc4ba01-323f-4bcb-9bc3-d639514f6088" containerName="init" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.546474 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cc4ba01-323f-4bcb-9bc3-d639514f6088" containerName="init" Feb 20 08:23:19 crc kubenswrapper[4948]: E0220 08:23:19.546494 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cc4ba01-323f-4bcb-9bc3-d639514f6088" containerName="dnsmasq-dns" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.546503 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cc4ba01-323f-4bcb-9bc3-d639514f6088" containerName="dnsmasq-dns" Feb 20 08:23:19 crc kubenswrapper[4948]: E0220 08:23:19.546522 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0661780-ca2b-42ae-be25-5d0e914fbb49" containerName="neutron-httpd" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.546529 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0661780-ca2b-42ae-be25-5d0e914fbb49" containerName="neutron-httpd" Feb 20 08:23:19 crc kubenswrapper[4948]: E0220 08:23:19.546546 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0661780-ca2b-42ae-be25-5d0e914fbb49" containerName="neutron-api" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.546557 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0661780-ca2b-42ae-be25-5d0e914fbb49" containerName="neutron-api" Feb 20 08:23:19 crc kubenswrapper[4948]: E0220 08:23:19.546579 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75294aee-c7f3-4b0b-82fc-df058490b90f" containerName="cinder-scheduler" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.546586 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="75294aee-c7f3-4b0b-82fc-df058490b90f" containerName="cinder-scheduler" Feb 20 08:23:19 crc kubenswrapper[4948]: E0220 08:23:19.546605 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75294aee-c7f3-4b0b-82fc-df058490b90f" containerName="probe" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.546612 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="75294aee-c7f3-4b0b-82fc-df058490b90f" containerName="probe" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.546815 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="75294aee-c7f3-4b0b-82fc-df058490b90f" containerName="probe" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.546843 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="75294aee-c7f3-4b0b-82fc-df058490b90f" containerName="cinder-scheduler" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.546860 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0661780-ca2b-42ae-be25-5d0e914fbb49" containerName="neutron-api" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.546878 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0661780-ca2b-42ae-be25-5d0e914fbb49" containerName="neutron-httpd" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.546892 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cc4ba01-323f-4bcb-9bc3-d639514f6088" containerName="dnsmasq-dns" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.548288 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.551000 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.555845 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.571985 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.572577 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.572625 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-config-data\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.572677 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8w4r\" (UniqueName: \"kubernetes.io/projected/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-kube-api-access-m8w4r\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.572765 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.572795 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-scripts\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.625069 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-86887846d-pqjcb"] Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.626362 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.638807 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-86887846d-pqjcb"] Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.674574 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.674640 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5047ff00-845b-459f-ac89-6cbf997f93fb-config-data\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.674676 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dsfr\" (UniqueName: \"kubernetes.io/projected/5047ff00-845b-459f-ac89-6cbf997f93fb-kube-api-access-4dsfr\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.674697 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.674717 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5047ff00-845b-459f-ac89-6cbf997f93fb-scripts\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.674737 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-config-data\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.674766 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8w4r\" (UniqueName: \"kubernetes.io/projected/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-kube-api-access-m8w4r\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.674780 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5047ff00-845b-459f-ac89-6cbf997f93fb-combined-ca-bundle\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.674808 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5047ff00-845b-459f-ac89-6cbf997f93fb-public-tls-certs\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.674849 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.674867 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5047ff00-845b-459f-ac89-6cbf997f93fb-internal-tls-certs\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.674884 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-scripts\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.674939 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5047ff00-845b-459f-ac89-6cbf997f93fb-logs\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.675049 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.683637 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.684390 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-config-data\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.684450 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-scripts\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.684454 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.695449 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8w4r\" (UniqueName: \"kubernetes.io/projected/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-kube-api-access-m8w4r\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.730952 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cc4ba01-323f-4bcb-9bc3-d639514f6088" path="/var/lib/kubelet/pods/0cc4ba01-323f-4bcb-9bc3-d639514f6088/volumes" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.731684 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75294aee-c7f3-4b0b-82fc-df058490b90f" path="/var/lib/kubelet/pods/75294aee-c7f3-4b0b-82fc-df058490b90f/volumes" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.776321 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5047ff00-845b-459f-ac89-6cbf997f93fb-combined-ca-bundle\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.776605 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5047ff00-845b-459f-ac89-6cbf997f93fb-public-tls-certs\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.776716 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5047ff00-845b-459f-ac89-6cbf997f93fb-internal-tls-certs\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.776826 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5047ff00-845b-459f-ac89-6cbf997f93fb-logs\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.776958 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5047ff00-845b-459f-ac89-6cbf997f93fb-config-data\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.777083 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dsfr\" (UniqueName: \"kubernetes.io/projected/5047ff00-845b-459f-ac89-6cbf997f93fb-kube-api-access-4dsfr\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.777157 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5047ff00-845b-459f-ac89-6cbf997f93fb-scripts\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.777750 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5047ff00-845b-459f-ac89-6cbf997f93fb-logs\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.779944 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5047ff00-845b-459f-ac89-6cbf997f93fb-combined-ca-bundle\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.780820 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5047ff00-845b-459f-ac89-6cbf997f93fb-internal-tls-certs\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.782878 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5047ff00-845b-459f-ac89-6cbf997f93fb-scripts\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.785273 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5047ff00-845b-459f-ac89-6cbf997f93fb-public-tls-certs\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.794253 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5047ff00-845b-459f-ac89-6cbf997f93fb-config-data\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.798108 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dsfr\" (UniqueName: \"kubernetes.io/projected/5047ff00-845b-459f-ac89-6cbf997f93fb-kube-api-access-4dsfr\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.873297 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.949428 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:20 crc kubenswrapper[4948]: I0220 08:23:20.442263 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 20 08:23:20 crc kubenswrapper[4948]: I0220 08:23:20.499067 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0","Type":"ContainerStarted","Data":"7b9175ec0f8b01106dc5de2e15b6e45b6d8ffb54de656411a7dfe063ac9da99b"} Feb 20 08:23:20 crc kubenswrapper[4948]: I0220 08:23:20.524872 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-86887846d-pqjcb"] Feb 20 08:23:21 crc kubenswrapper[4948]: I0220 08:23:21.517579 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0","Type":"ContainerStarted","Data":"9b3fe039662e55fa4b6cee25944a613b3149ab59c484a7308119f066ce7b1370"} Feb 20 08:23:21 crc kubenswrapper[4948]: I0220 08:23:21.520574 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-86887846d-pqjcb" event={"ID":"5047ff00-845b-459f-ac89-6cbf997f93fb","Type":"ContainerStarted","Data":"8d22e37d5f9b6a5c86cc676fc0e1a18b6aef0379685fd530b8da7ae8302c562a"} Feb 20 08:23:21 crc kubenswrapper[4948]: I0220 08:23:21.520602 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-86887846d-pqjcb" event={"ID":"5047ff00-845b-459f-ac89-6cbf997f93fb","Type":"ContainerStarted","Data":"4d24ae4e85b277b60efc36629974351db737751a376318ecdaabb0c430d33103"} Feb 20 08:23:21 crc kubenswrapper[4948]: I0220 08:23:21.520611 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-86887846d-pqjcb" event={"ID":"5047ff00-845b-459f-ac89-6cbf997f93fb","Type":"ContainerStarted","Data":"d25eb689b81e5ffd366d099059c81fcabd95ba9c317c99a3b01984365a662368"} Feb 20 08:23:21 crc kubenswrapper[4948]: I0220 08:23:21.520824 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:21 crc kubenswrapper[4948]: I0220 08:23:21.520869 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:21 crc kubenswrapper[4948]: I0220 08:23:21.548462 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-86887846d-pqjcb" podStartSLOduration=2.548445226 podStartE2EDuration="2.548445226s" podCreationTimestamp="2026-02-20 08:23:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:23:21.540704465 +0000 UTC m=+1050.515199285" watchObservedRunningTime="2026-02-20 08:23:21.548445226 +0000 UTC m=+1050.522940046" Feb 20 08:23:22 crc kubenswrapper[4948]: I0220 08:23:22.528877 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0","Type":"ContainerStarted","Data":"f35ed6518f2bf61623327f1157db4180ec50f47b3db49dac1d2fbf7c882e31df"} Feb 20 08:23:22 crc kubenswrapper[4948]: I0220 08:23:22.558108 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.558093732 podStartE2EDuration="3.558093732s" podCreationTimestamp="2026-02-20 08:23:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:23:22.555861157 +0000 UTC m=+1051.530355977" watchObservedRunningTime="2026-02-20 08:23:22.558093732 +0000 UTC m=+1051.532588552" Feb 20 08:23:22 crc kubenswrapper[4948]: I0220 08:23:22.890483 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.500845 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.502130 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.503574 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-4bss6" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.504573 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.517862 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.539510 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.548918 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq46g\" (UniqueName: \"kubernetes.io/projected/b2a76f68-3b4b-4993-aa4f-3b31b4ee1008-kube-api-access-sq46g\") pod \"openstackclient\" (UID: \"b2a76f68-3b4b-4993-aa4f-3b31b4ee1008\") " pod="openstack/openstackclient" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.549184 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b2a76f68-3b4b-4993-aa4f-3b31b4ee1008-openstack-config\") pod \"openstackclient\" (UID: \"b2a76f68-3b4b-4993-aa4f-3b31b4ee1008\") " pod="openstack/openstackclient" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.549280 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2a76f68-3b4b-4993-aa4f-3b31b4ee1008-combined-ca-bundle\") pod \"openstackclient\" (UID: \"b2a76f68-3b4b-4993-aa4f-3b31b4ee1008\") " pod="openstack/openstackclient" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.549366 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b2a76f68-3b4b-4993-aa4f-3b31b4ee1008-openstack-config-secret\") pod \"openstackclient\" (UID: \"b2a76f68-3b4b-4993-aa4f-3b31b4ee1008\") " pod="openstack/openstackclient" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.651723 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b2a76f68-3b4b-4993-aa4f-3b31b4ee1008-openstack-config\") pod \"openstackclient\" (UID: \"b2a76f68-3b4b-4993-aa4f-3b31b4ee1008\") " pod="openstack/openstackclient" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.651780 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2a76f68-3b4b-4993-aa4f-3b31b4ee1008-combined-ca-bundle\") pod \"openstackclient\" (UID: \"b2a76f68-3b4b-4993-aa4f-3b31b4ee1008\") " pod="openstack/openstackclient" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.651809 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b2a76f68-3b4b-4993-aa4f-3b31b4ee1008-openstack-config-secret\") pod \"openstackclient\" (UID: \"b2a76f68-3b4b-4993-aa4f-3b31b4ee1008\") " pod="openstack/openstackclient" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.652605 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b2a76f68-3b4b-4993-aa4f-3b31b4ee1008-openstack-config\") pod \"openstackclient\" (UID: \"b2a76f68-3b4b-4993-aa4f-3b31b4ee1008\") " pod="openstack/openstackclient" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.653084 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq46g\" (UniqueName: \"kubernetes.io/projected/b2a76f68-3b4b-4993-aa4f-3b31b4ee1008-kube-api-access-sq46g\") pod \"openstackclient\" (UID: \"b2a76f68-3b4b-4993-aa4f-3b31b4ee1008\") " pod="openstack/openstackclient" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.666060 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2a76f68-3b4b-4993-aa4f-3b31b4ee1008-combined-ca-bundle\") pod \"openstackclient\" (UID: \"b2a76f68-3b4b-4993-aa4f-3b31b4ee1008\") " pod="openstack/openstackclient" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.666089 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b2a76f68-3b4b-4993-aa4f-3b31b4ee1008-openstack-config-secret\") pod \"openstackclient\" (UID: \"b2a76f68-3b4b-4993-aa4f-3b31b4ee1008\") " pod="openstack/openstackclient" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.696072 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq46g\" (UniqueName: \"kubernetes.io/projected/b2a76f68-3b4b-4993-aa4f-3b31b4ee1008-kube-api-access-sq46g\") pod \"openstackclient\" (UID: \"b2a76f68-3b4b-4993-aa4f-3b31b4ee1008\") " pod="openstack/openstackclient" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.857156 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.898027 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Feb 20 08:23:24 crc kubenswrapper[4948]: I0220 08:23:24.204576 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:24 crc kubenswrapper[4948]: I0220 08:23:24.205682 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:24 crc kubenswrapper[4948]: I0220 08:23:24.299151 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-65487cc4d6-7gdvr"] Feb 20 08:23:24 crc kubenswrapper[4948]: I0220 08:23:24.300321 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-65487cc4d6-7gdvr" podUID="6a669dd0-d287-4d07-9eec-aea5ed80c791" containerName="barbican-api" containerID="cri-o://ab05c86cbd431ad2cc4a3cef37a6fc481837d1b6073c9a80ed0ee92d983240c8" gracePeriod=30 Feb 20 08:23:24 crc kubenswrapper[4948]: I0220 08:23:24.299935 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-65487cc4d6-7gdvr" podUID="6a669dd0-d287-4d07-9eec-aea5ed80c791" containerName="barbican-api-log" containerID="cri-o://86e858c136de80a807f997b8991e918a47db690a2c7b9fcc3401caa707f989dd" gracePeriod=30 Feb 20 08:23:24 crc kubenswrapper[4948]: I0220 08:23:24.450195 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 20 08:23:24 crc kubenswrapper[4948]: W0220 08:23:24.466296 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb2a76f68_3b4b_4993_aa4f_3b31b4ee1008.slice/crio-2a29f53c784327d73782ce53c1e38b1da346ef1d3769ab76fd235912c07e2a35 WatchSource:0}: Error finding container 2a29f53c784327d73782ce53c1e38b1da346ef1d3769ab76fd235912c07e2a35: Status 404 returned error can't find the container with id 2a29f53c784327d73782ce53c1e38b1da346ef1d3769ab76fd235912c07e2a35 Feb 20 08:23:24 crc kubenswrapper[4948]: I0220 08:23:24.558542 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"b2a76f68-3b4b-4993-aa4f-3b31b4ee1008","Type":"ContainerStarted","Data":"2a29f53c784327d73782ce53c1e38b1da346ef1d3769ab76fd235912c07e2a35"} Feb 20 08:23:24 crc kubenswrapper[4948]: I0220 08:23:24.560582 4948 generic.go:334] "Generic (PLEG): container finished" podID="6a669dd0-d287-4d07-9eec-aea5ed80c791" containerID="86e858c136de80a807f997b8991e918a47db690a2c7b9fcc3401caa707f989dd" exitCode=143 Feb 20 08:23:24 crc kubenswrapper[4948]: I0220 08:23:24.560633 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65487cc4d6-7gdvr" event={"ID":"6a669dd0-d287-4d07-9eec-aea5ed80c791","Type":"ContainerDied","Data":"86e858c136de80a807f997b8991e918a47db690a2c7b9fcc3401caa707f989dd"} Feb 20 08:23:24 crc kubenswrapper[4948]: I0220 08:23:24.873961 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 20 08:23:27 crc kubenswrapper[4948]: I0220 08:23:27.466601 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-65487cc4d6-7gdvr" podUID="6a669dd0-d287-4d07-9eec-aea5ed80c791" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.166:9311/healthcheck\": read tcp 10.217.0.2:48370->10.217.0.166:9311: read: connection reset by peer" Feb 20 08:23:27 crc kubenswrapper[4948]: I0220 08:23:27.466679 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-65487cc4d6-7gdvr" podUID="6a669dd0-d287-4d07-9eec-aea5ed80c791" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.166:9311/healthcheck\": read tcp 10.217.0.2:48382->10.217.0.166:9311: read: connection reset by peer" Feb 20 08:23:27 crc kubenswrapper[4948]: I0220 08:23:27.591152 4948 generic.go:334] "Generic (PLEG): container finished" podID="6a669dd0-d287-4d07-9eec-aea5ed80c791" containerID="ab05c86cbd431ad2cc4a3cef37a6fc481837d1b6073c9a80ed0ee92d983240c8" exitCode=0 Feb 20 08:23:27 crc kubenswrapper[4948]: I0220 08:23:27.591364 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65487cc4d6-7gdvr" event={"ID":"6a669dd0-d287-4d07-9eec-aea5ed80c791","Type":"ContainerDied","Data":"ab05c86cbd431ad2cc4a3cef37a6fc481837d1b6073c9a80ed0ee92d983240c8"} Feb 20 08:23:27 crc kubenswrapper[4948]: I0220 08:23:27.947945 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.052082 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-combined-ca-bundle\") pod \"6a669dd0-d287-4d07-9eec-aea5ed80c791\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.052210 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-config-data\") pod \"6a669dd0-d287-4d07-9eec-aea5ed80c791\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.052280 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-config-data-custom\") pod \"6a669dd0-d287-4d07-9eec-aea5ed80c791\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.052379 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ld59l\" (UniqueName: \"kubernetes.io/projected/6a669dd0-d287-4d07-9eec-aea5ed80c791-kube-api-access-ld59l\") pod \"6a669dd0-d287-4d07-9eec-aea5ed80c791\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.052423 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a669dd0-d287-4d07-9eec-aea5ed80c791-logs\") pod \"6a669dd0-d287-4d07-9eec-aea5ed80c791\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.053144 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a669dd0-d287-4d07-9eec-aea5ed80c791-logs" (OuterVolumeSpecName: "logs") pod "6a669dd0-d287-4d07-9eec-aea5ed80c791" (UID: "6a669dd0-d287-4d07-9eec-aea5ed80c791"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.057624 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a669dd0-d287-4d07-9eec-aea5ed80c791-kube-api-access-ld59l" (OuterVolumeSpecName: "kube-api-access-ld59l") pod "6a669dd0-d287-4d07-9eec-aea5ed80c791" (UID: "6a669dd0-d287-4d07-9eec-aea5ed80c791"). InnerVolumeSpecName "kube-api-access-ld59l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.089875 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6a669dd0-d287-4d07-9eec-aea5ed80c791" (UID: "6a669dd0-d287-4d07-9eec-aea5ed80c791"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.110449 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6a669dd0-d287-4d07-9eec-aea5ed80c791" (UID: "6a669dd0-d287-4d07-9eec-aea5ed80c791"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.129182 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-config-data" (OuterVolumeSpecName: "config-data") pod "6a669dd0-d287-4d07-9eec-aea5ed80c791" (UID: "6a669dd0-d287-4d07-9eec-aea5ed80c791"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.154245 4948 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.154274 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ld59l\" (UniqueName: \"kubernetes.io/projected/6a669dd0-d287-4d07-9eec-aea5ed80c791-kube-api-access-ld59l\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.154286 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a669dd0-d287-4d07-9eec-aea5ed80c791-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.154296 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.154305 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.272184 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.272492 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="44092649-be82-4225-8cfc-d37dbe840877" containerName="proxy-httpd" containerID="cri-o://e85736cf9a602b496a8372a3fe185272b09bf09e11882adec2e63f62c62bf3d9" gracePeriod=30 Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.272584 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="44092649-be82-4225-8cfc-d37dbe840877" containerName="sg-core" containerID="cri-o://d829cc037c7ba0bd0abfe863863495a3388592344985fae656879df8c40ec990" gracePeriod=30 Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.272789 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="44092649-be82-4225-8cfc-d37dbe840877" containerName="ceilometer-notification-agent" containerID="cri-o://7e6d02d6234c2e72678f463fac466a22ef0548ef1c161d0d356079f01463ff09" gracePeriod=30 Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.272859 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="44092649-be82-4225-8cfc-d37dbe840877" containerName="ceilometer-central-agent" containerID="cri-o://eab93963947d8aecd2b45387be911464a6576e8236c9ccd3d0a009ad65054a74" gracePeriod=30 Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.289287 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="44092649-be82-4225-8cfc-d37dbe840877" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.167:3000/\": EOF" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.412126 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-x2pm7"] Feb 20 08:23:28 crc kubenswrapper[4948]: E0220 08:23:28.412847 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a669dd0-d287-4d07-9eec-aea5ed80c791" containerName="barbican-api-log" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.412868 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a669dd0-d287-4d07-9eec-aea5ed80c791" containerName="barbican-api-log" Feb 20 08:23:28 crc kubenswrapper[4948]: E0220 08:23:28.412884 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a669dd0-d287-4d07-9eec-aea5ed80c791" containerName="barbican-api" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.412891 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a669dd0-d287-4d07-9eec-aea5ed80c791" containerName="barbican-api" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.413083 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a669dd0-d287-4d07-9eec-aea5ed80c791" containerName="barbican-api" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.413102 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a669dd0-d287-4d07-9eec-aea5ed80c791" containerName="barbican-api-log" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.413658 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-x2pm7" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.434544 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-x2pm7"] Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.460709 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa398740-9a54-472f-9736-0959dd54b657-operator-scripts\") pod \"nova-api-db-create-x2pm7\" (UID: \"fa398740-9a54-472f-9736-0959dd54b657\") " pod="openstack/nova-api-db-create-x2pm7" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.460777 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rx99\" (UniqueName: \"kubernetes.io/projected/fa398740-9a54-472f-9736-0959dd54b657-kube-api-access-7rx99\") pod \"nova-api-db-create-x2pm7\" (UID: \"fa398740-9a54-472f-9736-0959dd54b657\") " pod="openstack/nova-api-db-create-x2pm7" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.562552 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rx99\" (UniqueName: \"kubernetes.io/projected/fa398740-9a54-472f-9736-0959dd54b657-kube-api-access-7rx99\") pod \"nova-api-db-create-x2pm7\" (UID: \"fa398740-9a54-472f-9736-0959dd54b657\") " pod="openstack/nova-api-db-create-x2pm7" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.562761 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa398740-9a54-472f-9736-0959dd54b657-operator-scripts\") pod \"nova-api-db-create-x2pm7\" (UID: \"fa398740-9a54-472f-9736-0959dd54b657\") " pod="openstack/nova-api-db-create-x2pm7" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.563808 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa398740-9a54-472f-9736-0959dd54b657-operator-scripts\") pod \"nova-api-db-create-x2pm7\" (UID: \"fa398740-9a54-472f-9736-0959dd54b657\") " pod="openstack/nova-api-db-create-x2pm7" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.592836 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rx99\" (UniqueName: \"kubernetes.io/projected/fa398740-9a54-472f-9736-0959dd54b657-kube-api-access-7rx99\") pod \"nova-api-db-create-x2pm7\" (UID: \"fa398740-9a54-472f-9736-0959dd54b657\") " pod="openstack/nova-api-db-create-x2pm7" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.612829 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65487cc4d6-7gdvr" event={"ID":"6a669dd0-d287-4d07-9eec-aea5ed80c791","Type":"ContainerDied","Data":"b3b840325250a810c158ca1f61d60db618feec352dbf68e16aad14e15139327d"} Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.612888 4948 scope.go:117] "RemoveContainer" containerID="ab05c86cbd431ad2cc4a3cef37a6fc481837d1b6073c9a80ed0ee92d983240c8" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.613030 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.630532 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-7c7cd9876c-262qf"] Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.635311 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.642473 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.642732 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.643406 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.644153 4948 generic.go:334] "Generic (PLEG): container finished" podID="44092649-be82-4225-8cfc-d37dbe840877" containerID="e85736cf9a602b496a8372a3fe185272b09bf09e11882adec2e63f62c62bf3d9" exitCode=0 Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.644172 4948 generic.go:334] "Generic (PLEG): container finished" podID="44092649-be82-4225-8cfc-d37dbe840877" containerID="d829cc037c7ba0bd0abfe863863495a3388592344985fae656879df8c40ec990" exitCode=2 Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.644190 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"44092649-be82-4225-8cfc-d37dbe840877","Type":"ContainerDied","Data":"e85736cf9a602b496a8372a3fe185272b09bf09e11882adec2e63f62c62bf3d9"} Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.644212 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"44092649-be82-4225-8cfc-d37dbe840877","Type":"ContainerDied","Data":"d829cc037c7ba0bd0abfe863863495a3388592344985fae656879df8c40ec990"} Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.664161 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30f48966-e44d-44bc-a868-7e8119ac186c-config-data\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.664219 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30f48966-e44d-44bc-a868-7e8119ac186c-combined-ca-bundle\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.664699 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgtbs\" (UniqueName: \"kubernetes.io/projected/30f48966-e44d-44bc-a868-7e8119ac186c-kube-api-access-lgtbs\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.664753 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30f48966-e44d-44bc-a868-7e8119ac186c-run-httpd\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.664769 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/30f48966-e44d-44bc-a868-7e8119ac186c-public-tls-certs\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.664788 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/30f48966-e44d-44bc-a868-7e8119ac186c-etc-swift\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.664835 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/30f48966-e44d-44bc-a868-7e8119ac186c-internal-tls-certs\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.664850 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30f48966-e44d-44bc-a868-7e8119ac186c-log-httpd\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.670068 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-mgnlb"] Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.672926 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-mgnlb" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.701208 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-mgnlb"] Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.716811 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7c7cd9876c-262qf"] Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.728204 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-aecf-account-create-update-gnnvg"] Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.729603 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-aecf-account-create-update-gnnvg" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.731879 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.747945 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-aecf-account-create-update-gnnvg"] Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.750196 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-x2pm7" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.755530 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-65487cc4d6-7gdvr"] Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.765995 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-65487cc4d6-7gdvr"] Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.766202 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/30f48966-e44d-44bc-a868-7e8119ac186c-internal-tls-certs\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.766253 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20b61242-b8c1-4e1b-b319-3eafb79ef048-operator-scripts\") pod \"nova-cell0-db-create-mgnlb\" (UID: \"20b61242-b8c1-4e1b-b319-3eafb79ef048\") " pod="openstack/nova-cell0-db-create-mgnlb" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.766283 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30f48966-e44d-44bc-a868-7e8119ac186c-log-httpd\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.766366 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30f48966-e44d-44bc-a868-7e8119ac186c-config-data\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.766417 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a311ded-ab56-4f07-8b7f-9dcc4d70f647-operator-scripts\") pod \"nova-api-aecf-account-create-update-gnnvg\" (UID: \"8a311ded-ab56-4f07-8b7f-9dcc4d70f647\") " pod="openstack/nova-api-aecf-account-create-update-gnnvg" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.766457 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30f48966-e44d-44bc-a868-7e8119ac186c-combined-ca-bundle\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.766490 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgtbs\" (UniqueName: \"kubernetes.io/projected/30f48966-e44d-44bc-a868-7e8119ac186c-kube-api-access-lgtbs\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.766554 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30f48966-e44d-44bc-a868-7e8119ac186c-run-httpd\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.766579 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/30f48966-e44d-44bc-a868-7e8119ac186c-public-tls-certs\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.766604 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/30f48966-e44d-44bc-a868-7e8119ac186c-etc-swift\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.766643 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ssjs\" (UniqueName: \"kubernetes.io/projected/20b61242-b8c1-4e1b-b319-3eafb79ef048-kube-api-access-2ssjs\") pod \"nova-cell0-db-create-mgnlb\" (UID: \"20b61242-b8c1-4e1b-b319-3eafb79ef048\") " pod="openstack/nova-cell0-db-create-mgnlb" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.766679 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ttpc\" (UniqueName: \"kubernetes.io/projected/8a311ded-ab56-4f07-8b7f-9dcc4d70f647-kube-api-access-7ttpc\") pod \"nova-api-aecf-account-create-update-gnnvg\" (UID: \"8a311ded-ab56-4f07-8b7f-9dcc4d70f647\") " pod="openstack/nova-api-aecf-account-create-update-gnnvg" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.766838 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30f48966-e44d-44bc-a868-7e8119ac186c-log-httpd\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.767134 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30f48966-e44d-44bc-a868-7e8119ac186c-run-httpd\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.771457 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/30f48966-e44d-44bc-a868-7e8119ac186c-internal-tls-certs\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.772067 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/30f48966-e44d-44bc-a868-7e8119ac186c-etc-swift\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.783866 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/30f48966-e44d-44bc-a868-7e8119ac186c-public-tls-certs\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.784381 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30f48966-e44d-44bc-a868-7e8119ac186c-combined-ca-bundle\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.788394 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30f48966-e44d-44bc-a868-7e8119ac186c-config-data\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.788662 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgtbs\" (UniqueName: \"kubernetes.io/projected/30f48966-e44d-44bc-a868-7e8119ac186c-kube-api-access-lgtbs\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.798469 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-2rgc6"] Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.807620 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-2rgc6" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.821167 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-2rgc6"] Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.841490 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-78de-account-create-update-fwjdc"] Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.842805 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-78de-account-create-update-fwjdc" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.844418 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.849636 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-78de-account-create-update-fwjdc"] Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.867365 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20b61242-b8c1-4e1b-b319-3eafb79ef048-operator-scripts\") pod \"nova-cell0-db-create-mgnlb\" (UID: \"20b61242-b8c1-4e1b-b319-3eafb79ef048\") " pod="openstack/nova-cell0-db-create-mgnlb" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.867476 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpq6v\" (UniqueName: \"kubernetes.io/projected/8b167b56-655a-4220-9823-8cc606f5f034-kube-api-access-fpq6v\") pod \"nova-cell0-78de-account-create-update-fwjdc\" (UID: \"8b167b56-655a-4220-9823-8cc606f5f034\") " pod="openstack/nova-cell0-78de-account-create-update-fwjdc" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.867535 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hgw6\" (UniqueName: \"kubernetes.io/projected/a5cb8c15-68fc-47db-9f68-1a8401f99f88-kube-api-access-2hgw6\") pod \"nova-cell1-db-create-2rgc6\" (UID: \"a5cb8c15-68fc-47db-9f68-1a8401f99f88\") " pod="openstack/nova-cell1-db-create-2rgc6" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.867556 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a311ded-ab56-4f07-8b7f-9dcc4d70f647-operator-scripts\") pod \"nova-api-aecf-account-create-update-gnnvg\" (UID: \"8a311ded-ab56-4f07-8b7f-9dcc4d70f647\") " pod="openstack/nova-api-aecf-account-create-update-gnnvg" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.867609 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5cb8c15-68fc-47db-9f68-1a8401f99f88-operator-scripts\") pod \"nova-cell1-db-create-2rgc6\" (UID: \"a5cb8c15-68fc-47db-9f68-1a8401f99f88\") " pod="openstack/nova-cell1-db-create-2rgc6" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.867655 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ssjs\" (UniqueName: \"kubernetes.io/projected/20b61242-b8c1-4e1b-b319-3eafb79ef048-kube-api-access-2ssjs\") pod \"nova-cell0-db-create-mgnlb\" (UID: \"20b61242-b8c1-4e1b-b319-3eafb79ef048\") " pod="openstack/nova-cell0-db-create-mgnlb" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.867681 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b167b56-655a-4220-9823-8cc606f5f034-operator-scripts\") pod \"nova-cell0-78de-account-create-update-fwjdc\" (UID: \"8b167b56-655a-4220-9823-8cc606f5f034\") " pod="openstack/nova-cell0-78de-account-create-update-fwjdc" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.867699 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ttpc\" (UniqueName: \"kubernetes.io/projected/8a311ded-ab56-4f07-8b7f-9dcc4d70f647-kube-api-access-7ttpc\") pod \"nova-api-aecf-account-create-update-gnnvg\" (UID: \"8a311ded-ab56-4f07-8b7f-9dcc4d70f647\") " pod="openstack/nova-api-aecf-account-create-update-gnnvg" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.868245 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20b61242-b8c1-4e1b-b319-3eafb79ef048-operator-scripts\") pod \"nova-cell0-db-create-mgnlb\" (UID: \"20b61242-b8c1-4e1b-b319-3eafb79ef048\") " pod="openstack/nova-cell0-db-create-mgnlb" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.868584 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a311ded-ab56-4f07-8b7f-9dcc4d70f647-operator-scripts\") pod \"nova-api-aecf-account-create-update-gnnvg\" (UID: \"8a311ded-ab56-4f07-8b7f-9dcc4d70f647\") " pod="openstack/nova-api-aecf-account-create-update-gnnvg" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.886963 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ttpc\" (UniqueName: \"kubernetes.io/projected/8a311ded-ab56-4f07-8b7f-9dcc4d70f647-kube-api-access-7ttpc\") pod \"nova-api-aecf-account-create-update-gnnvg\" (UID: \"8a311ded-ab56-4f07-8b7f-9dcc4d70f647\") " pod="openstack/nova-api-aecf-account-create-update-gnnvg" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.894728 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ssjs\" (UniqueName: \"kubernetes.io/projected/20b61242-b8c1-4e1b-b319-3eafb79ef048-kube-api-access-2ssjs\") pod \"nova-cell0-db-create-mgnlb\" (UID: \"20b61242-b8c1-4e1b-b319-3eafb79ef048\") " pod="openstack/nova-cell0-db-create-mgnlb" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.956731 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.970658 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b167b56-655a-4220-9823-8cc606f5f034-operator-scripts\") pod \"nova-cell0-78de-account-create-update-fwjdc\" (UID: \"8b167b56-655a-4220-9823-8cc606f5f034\") " pod="openstack/nova-cell0-78de-account-create-update-fwjdc" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.971498 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b167b56-655a-4220-9823-8cc606f5f034-operator-scripts\") pod \"nova-cell0-78de-account-create-update-fwjdc\" (UID: \"8b167b56-655a-4220-9823-8cc606f5f034\") " pod="openstack/nova-cell0-78de-account-create-update-fwjdc" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.971944 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpq6v\" (UniqueName: \"kubernetes.io/projected/8b167b56-655a-4220-9823-8cc606f5f034-kube-api-access-fpq6v\") pod \"nova-cell0-78de-account-create-update-fwjdc\" (UID: \"8b167b56-655a-4220-9823-8cc606f5f034\") " pod="openstack/nova-cell0-78de-account-create-update-fwjdc" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.972841 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hgw6\" (UniqueName: \"kubernetes.io/projected/a5cb8c15-68fc-47db-9f68-1a8401f99f88-kube-api-access-2hgw6\") pod \"nova-cell1-db-create-2rgc6\" (UID: \"a5cb8c15-68fc-47db-9f68-1a8401f99f88\") " pod="openstack/nova-cell1-db-create-2rgc6" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.973267 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5cb8c15-68fc-47db-9f68-1a8401f99f88-operator-scripts\") pod \"nova-cell1-db-create-2rgc6\" (UID: \"a5cb8c15-68fc-47db-9f68-1a8401f99f88\") " pod="openstack/nova-cell1-db-create-2rgc6" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.975287 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5cb8c15-68fc-47db-9f68-1a8401f99f88-operator-scripts\") pod \"nova-cell1-db-create-2rgc6\" (UID: \"a5cb8c15-68fc-47db-9f68-1a8401f99f88\") " pod="openstack/nova-cell1-db-create-2rgc6" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.988732 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-mgnlb" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.988951 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpq6v\" (UniqueName: \"kubernetes.io/projected/8b167b56-655a-4220-9823-8cc606f5f034-kube-api-access-fpq6v\") pod \"nova-cell0-78de-account-create-update-fwjdc\" (UID: \"8b167b56-655a-4220-9823-8cc606f5f034\") " pod="openstack/nova-cell0-78de-account-create-update-fwjdc" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.996454 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hgw6\" (UniqueName: \"kubernetes.io/projected/a5cb8c15-68fc-47db-9f68-1a8401f99f88-kube-api-access-2hgw6\") pod \"nova-cell1-db-create-2rgc6\" (UID: \"a5cb8c15-68fc-47db-9f68-1a8401f99f88\") " pod="openstack/nova-cell1-db-create-2rgc6" Feb 20 08:23:29 crc kubenswrapper[4948]: I0220 08:23:29.028120 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-ca33-account-create-update-dwqpd"] Feb 20 08:23:29 crc kubenswrapper[4948]: I0220 08:23:29.029672 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ca33-account-create-update-dwqpd" Feb 20 08:23:29 crc kubenswrapper[4948]: I0220 08:23:29.031681 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Feb 20 08:23:29 crc kubenswrapper[4948]: I0220 08:23:29.039070 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-ca33-account-create-update-dwqpd"] Feb 20 08:23:29 crc kubenswrapper[4948]: I0220 08:23:29.051621 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-aecf-account-create-update-gnnvg" Feb 20 08:23:29 crc kubenswrapper[4948]: I0220 08:23:29.177280 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f-operator-scripts\") pod \"nova-cell1-ca33-account-create-update-dwqpd\" (UID: \"b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f\") " pod="openstack/nova-cell1-ca33-account-create-update-dwqpd" Feb 20 08:23:29 crc kubenswrapper[4948]: I0220 08:23:29.177364 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vl2t2\" (UniqueName: \"kubernetes.io/projected/b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f-kube-api-access-vl2t2\") pod \"nova-cell1-ca33-account-create-update-dwqpd\" (UID: \"b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f\") " pod="openstack/nova-cell1-ca33-account-create-update-dwqpd" Feb 20 08:23:29 crc kubenswrapper[4948]: I0220 08:23:29.188277 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-2rgc6" Feb 20 08:23:29 crc kubenswrapper[4948]: I0220 08:23:29.237665 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-78de-account-create-update-fwjdc" Feb 20 08:23:29 crc kubenswrapper[4948]: I0220 08:23:29.279026 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f-operator-scripts\") pod \"nova-cell1-ca33-account-create-update-dwqpd\" (UID: \"b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f\") " pod="openstack/nova-cell1-ca33-account-create-update-dwqpd" Feb 20 08:23:29 crc kubenswrapper[4948]: I0220 08:23:29.279119 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vl2t2\" (UniqueName: \"kubernetes.io/projected/b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f-kube-api-access-vl2t2\") pod \"nova-cell1-ca33-account-create-update-dwqpd\" (UID: \"b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f\") " pod="openstack/nova-cell1-ca33-account-create-update-dwqpd" Feb 20 08:23:29 crc kubenswrapper[4948]: I0220 08:23:29.279921 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f-operator-scripts\") pod \"nova-cell1-ca33-account-create-update-dwqpd\" (UID: \"b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f\") " pod="openstack/nova-cell1-ca33-account-create-update-dwqpd" Feb 20 08:23:29 crc kubenswrapper[4948]: I0220 08:23:29.297404 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vl2t2\" (UniqueName: \"kubernetes.io/projected/b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f-kube-api-access-vl2t2\") pod \"nova-cell1-ca33-account-create-update-dwqpd\" (UID: \"b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f\") " pod="openstack/nova-cell1-ca33-account-create-update-dwqpd" Feb 20 08:23:29 crc kubenswrapper[4948]: I0220 08:23:29.381928 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ca33-account-create-update-dwqpd" Feb 20 08:23:29 crc kubenswrapper[4948]: I0220 08:23:29.657247 4948 generic.go:334] "Generic (PLEG): container finished" podID="44092649-be82-4225-8cfc-d37dbe840877" containerID="eab93963947d8aecd2b45387be911464a6576e8236c9ccd3d0a009ad65054a74" exitCode=0 Feb 20 08:23:29 crc kubenswrapper[4948]: I0220 08:23:29.657284 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"44092649-be82-4225-8cfc-d37dbe840877","Type":"ContainerDied","Data":"eab93963947d8aecd2b45387be911464a6576e8236c9ccd3d0a009ad65054a74"} Feb 20 08:23:29 crc kubenswrapper[4948]: I0220 08:23:29.735961 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a669dd0-d287-4d07-9eec-aea5ed80c791" path="/var/lib/kubelet/pods/6a669dd0-d287-4d07-9eec-aea5ed80c791/volumes" Feb 20 08:23:30 crc kubenswrapper[4948]: I0220 08:23:30.092399 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 20 08:23:31 crc kubenswrapper[4948]: I0220 08:23:31.678460 4948 generic.go:334] "Generic (PLEG): container finished" podID="44092649-be82-4225-8cfc-d37dbe840877" containerID="7e6d02d6234c2e72678f463fac466a22ef0548ef1c161d0d356079f01463ff09" exitCode=0 Feb 20 08:23:31 crc kubenswrapper[4948]: I0220 08:23:31.678535 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"44092649-be82-4225-8cfc-d37dbe840877","Type":"ContainerDied","Data":"7e6d02d6234c2e72678f463fac466a22ef0548ef1c161d0d356079f01463ff09"} Feb 20 08:23:33 crc kubenswrapper[4948]: I0220 08:23:33.870656 4948 scope.go:117] "RemoveContainer" containerID="86e858c136de80a807f997b8991e918a47db690a2c7b9fcc3401caa707f989dd" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.327023 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.470355 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-config-data\") pod \"44092649-be82-4225-8cfc-d37dbe840877\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.470402 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-sg-core-conf-yaml\") pod \"44092649-be82-4225-8cfc-d37dbe840877\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.470506 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-scripts\") pod \"44092649-be82-4225-8cfc-d37dbe840877\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.470526 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-combined-ca-bundle\") pod \"44092649-be82-4225-8cfc-d37dbe840877\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.470633 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/44092649-be82-4225-8cfc-d37dbe840877-log-httpd\") pod \"44092649-be82-4225-8cfc-d37dbe840877\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.470667 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rkd9\" (UniqueName: \"kubernetes.io/projected/44092649-be82-4225-8cfc-d37dbe840877-kube-api-access-4rkd9\") pod \"44092649-be82-4225-8cfc-d37dbe840877\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.470689 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/44092649-be82-4225-8cfc-d37dbe840877-run-httpd\") pod \"44092649-be82-4225-8cfc-d37dbe840877\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.471511 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44092649-be82-4225-8cfc-d37dbe840877-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "44092649-be82-4225-8cfc-d37dbe840877" (UID: "44092649-be82-4225-8cfc-d37dbe840877"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.472240 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44092649-be82-4225-8cfc-d37dbe840877-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "44092649-be82-4225-8cfc-d37dbe840877" (UID: "44092649-be82-4225-8cfc-d37dbe840877"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.477629 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44092649-be82-4225-8cfc-d37dbe840877-kube-api-access-4rkd9" (OuterVolumeSpecName: "kube-api-access-4rkd9") pod "44092649-be82-4225-8cfc-d37dbe840877" (UID: "44092649-be82-4225-8cfc-d37dbe840877"). InnerVolumeSpecName "kube-api-access-4rkd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.483036 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-scripts" (OuterVolumeSpecName: "scripts") pod "44092649-be82-4225-8cfc-d37dbe840877" (UID: "44092649-be82-4225-8cfc-d37dbe840877"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.506921 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "44092649-be82-4225-8cfc-d37dbe840877" (UID: "44092649-be82-4225-8cfc-d37dbe840877"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.558807 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "44092649-be82-4225-8cfc-d37dbe840877" (UID: "44092649-be82-4225-8cfc-d37dbe840877"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.572886 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rkd9\" (UniqueName: \"kubernetes.io/projected/44092649-be82-4225-8cfc-d37dbe840877-kube-api-access-4rkd9\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.572920 4948 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/44092649-be82-4225-8cfc-d37dbe840877-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.572929 4948 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.572941 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.572949 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.572957 4948 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/44092649-be82-4225-8cfc-d37dbe840877-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.603092 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-config-data" (OuterVolumeSpecName: "config-data") pod "44092649-be82-4225-8cfc-d37dbe840877" (UID: "44092649-be82-4225-8cfc-d37dbe840877"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.638309 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-2rgc6"] Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.649023 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-aecf-account-create-update-gnnvg"] Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.674869 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.717869 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"b2a76f68-3b4b-4993-aa4f-3b31b4ee1008","Type":"ContainerStarted","Data":"b1aae9b1fd753863bdd28b79973fb865f9bfe71025414314382214f9398365f4"} Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.720391 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-2rgc6" event={"ID":"a5cb8c15-68fc-47db-9f68-1a8401f99f88","Type":"ContainerStarted","Data":"7de096242d5ba0d916d6903858d689e5e9254091b9036d8255334ff0ea8f1f16"} Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.726821 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"44092649-be82-4225-8cfc-d37dbe840877","Type":"ContainerDied","Data":"8266937b599da958a185777fc1969453e86491578f18e8d8f35aa2e0681bba26"} Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.726829 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.726878 4948 scope.go:117] "RemoveContainer" containerID="e85736cf9a602b496a8372a3fe185272b09bf09e11882adec2e63f62c62bf3d9" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.730650 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-aecf-account-create-update-gnnvg" event={"ID":"8a311ded-ab56-4f07-8b7f-9dcc4d70f647","Type":"ContainerStarted","Data":"04a106975b06c46c735fa10604e40565f439a58073da7467ca0316db88eec4f9"} Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.740097 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.204848791 podStartE2EDuration="11.740078988s" podCreationTimestamp="2026-02-20 08:23:23 +0000 UTC" firstStartedPulling="2026-02-20 08:23:24.468637252 +0000 UTC m=+1053.443132072" lastFinishedPulling="2026-02-20 08:23:34.003867449 +0000 UTC m=+1062.978362269" observedRunningTime="2026-02-20 08:23:34.733116166 +0000 UTC m=+1063.707611006" watchObservedRunningTime="2026-02-20 08:23:34.740078988 +0000 UTC m=+1063.714573808" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.778152 4948 scope.go:117] "RemoveContainer" containerID="d829cc037c7ba0bd0abfe863863495a3388592344985fae656879df8c40ec990" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.779525 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-mgnlb"] Feb 20 08:23:34 crc kubenswrapper[4948]: W0220 08:23:34.781985 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20b61242_b8c1_4e1b_b319_3eafb79ef048.slice/crio-c5c2d7e3e5610297feff1dafc766462b7a8ac2739d1c5be80ff63c9dbf1bcc12 WatchSource:0}: Error finding container c5c2d7e3e5610297feff1dafc766462b7a8ac2739d1c5be80ff63c9dbf1bcc12: Status 404 returned error can't find the container with id c5c2d7e3e5610297feff1dafc766462b7a8ac2739d1c5be80ff63c9dbf1bcc12 Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.832098 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.851798 4948 scope.go:117] "RemoveContainer" containerID="7e6d02d6234c2e72678f463fac466a22ef0548ef1c161d0d356079f01463ff09" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.855589 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.876399 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:34 crc kubenswrapper[4948]: E0220 08:23:34.876840 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44092649-be82-4225-8cfc-d37dbe840877" containerName="ceilometer-central-agent" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.876856 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="44092649-be82-4225-8cfc-d37dbe840877" containerName="ceilometer-central-agent" Feb 20 08:23:34 crc kubenswrapper[4948]: E0220 08:23:34.876870 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44092649-be82-4225-8cfc-d37dbe840877" containerName="sg-core" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.876878 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="44092649-be82-4225-8cfc-d37dbe840877" containerName="sg-core" Feb 20 08:23:34 crc kubenswrapper[4948]: E0220 08:23:34.876908 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44092649-be82-4225-8cfc-d37dbe840877" containerName="proxy-httpd" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.876916 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="44092649-be82-4225-8cfc-d37dbe840877" containerName="proxy-httpd" Feb 20 08:23:34 crc kubenswrapper[4948]: E0220 08:23:34.876937 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44092649-be82-4225-8cfc-d37dbe840877" containerName="ceilometer-notification-agent" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.876944 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="44092649-be82-4225-8cfc-d37dbe840877" containerName="ceilometer-notification-agent" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.877163 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="44092649-be82-4225-8cfc-d37dbe840877" containerName="ceilometer-central-agent" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.877185 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="44092649-be82-4225-8cfc-d37dbe840877" containerName="sg-core" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.877193 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="44092649-be82-4225-8cfc-d37dbe840877" containerName="proxy-httpd" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.877215 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="44092649-be82-4225-8cfc-d37dbe840877" containerName="ceilometer-notification-agent" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.883695 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.885271 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.888341 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.888624 4948 scope.go:117] "RemoveContainer" containerID="eab93963947d8aecd2b45387be911464a6576e8236c9ccd3d0a009ad65054a74" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.888908 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.892989 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-78de-account-create-update-fwjdc"] Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.919168 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7c7cd9876c-262qf"] Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.929846 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-x2pm7"] Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.945999 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-ca33-account-create-update-dwqpd"] Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.081082 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.081184 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-run-httpd\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.081216 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-config-data\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.081400 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-scripts\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.081516 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.081626 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qc4df\" (UniqueName: \"kubernetes.io/projected/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-kube-api-access-qc4df\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.081681 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-log-httpd\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.182869 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.183049 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-run-httpd\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.183089 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-config-data\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.183128 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-scripts\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.183167 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.183194 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qc4df\" (UniqueName: \"kubernetes.io/projected/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-kube-api-access-qc4df\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.183232 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-log-httpd\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.183890 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-log-httpd\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.184013 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-run-httpd\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.188060 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-config-data\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.188886 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.189537 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-scripts\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.193556 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.202562 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qc4df\" (UniqueName: \"kubernetes.io/projected/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-kube-api-access-qc4df\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.363683 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.735106 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44092649-be82-4225-8cfc-d37dbe840877" path="/var/lib/kubelet/pods/44092649-be82-4225-8cfc-d37dbe840877/volumes" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.749878 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ca33-account-create-update-dwqpd" event={"ID":"b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f","Type":"ContainerStarted","Data":"97e6bc167799c61927aab7b946388571d89bdf96214e87d2ba307142c0a90dd5"} Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.749926 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ca33-account-create-update-dwqpd" event={"ID":"b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f","Type":"ContainerStarted","Data":"4a6359d724bfd5ae6bb40a902ffdabe96c01743e8c81b47b0e26ad235b7eaeff"} Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.753842 4948 generic.go:334] "Generic (PLEG): container finished" podID="8a311ded-ab56-4f07-8b7f-9dcc4d70f647" containerID="dd9bca0f8ae0d62df16de42592e7d1d0793a43cd05a5a5d5dcaf73199e707005" exitCode=0 Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.754218 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-aecf-account-create-update-gnnvg" event={"ID":"8a311ded-ab56-4f07-8b7f-9dcc4d70f647","Type":"ContainerDied","Data":"dd9bca0f8ae0d62df16de42592e7d1d0793a43cd05a5a5d5dcaf73199e707005"} Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.766317 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-x2pm7" event={"ID":"fa398740-9a54-472f-9736-0959dd54b657","Type":"ContainerStarted","Data":"be6725bc5e2da748cc92ca6d3d54cea0585e5ca878e37159db354dc55a1eae54"} Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.766521 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-x2pm7" event={"ID":"fa398740-9a54-472f-9736-0959dd54b657","Type":"ContainerStarted","Data":"372eccf775e290867aea35db34cd53f0b147aa6e61ff520ef1ac676e72e0326c"} Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.782137 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-ca33-account-create-update-dwqpd" podStartSLOduration=6.782116785 podStartE2EDuration="6.782116785s" podCreationTimestamp="2026-02-20 08:23:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:23:35.76371817 +0000 UTC m=+1064.738212990" watchObservedRunningTime="2026-02-20 08:23:35.782116785 +0000 UTC m=+1064.756611605" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.784845 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7c7cd9876c-262qf" event={"ID":"30f48966-e44d-44bc-a868-7e8119ac186c","Type":"ContainerStarted","Data":"811eeb0baa418ab2012f0916ae52adf1815521bd676db6d446531d2de318d57e"} Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.784992 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7c7cd9876c-262qf" event={"ID":"30f48966-e44d-44bc-a868-7e8119ac186c","Type":"ContainerStarted","Data":"5bc8f894cd28a290b6306bff2fa7f5125cf755eb2a8fdd1a4c96b9198eadef20"} Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.785071 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7c7cd9876c-262qf" event={"ID":"30f48966-e44d-44bc-a868-7e8119ac186c","Type":"ContainerStarted","Data":"5f5a0fa2e5552aae0ce9199499829b8bae2123b19a32b32e0c4140636cd5b2cf"} Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.785884 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.785986 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.802076 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-x2pm7" podStartSLOduration=7.802059238 podStartE2EDuration="7.802059238s" podCreationTimestamp="2026-02-20 08:23:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:23:35.801674869 +0000 UTC m=+1064.776169689" watchObservedRunningTime="2026-02-20 08:23:35.802059238 +0000 UTC m=+1064.776554058" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.806069 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-mgnlb" event={"ID":"20b61242-b8c1-4e1b-b319-3eafb79ef048","Type":"ContainerStarted","Data":"ca086fdde2d4c69122e9c700d4d4af3b8218c685639bf6df2d417cc08c1e51d3"} Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.806109 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-mgnlb" event={"ID":"20b61242-b8c1-4e1b-b319-3eafb79ef048","Type":"ContainerStarted","Data":"c5c2d7e3e5610297feff1dafc766462b7a8ac2739d1c5be80ff63c9dbf1bcc12"} Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.810395 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-78de-account-create-update-fwjdc" event={"ID":"8b167b56-655a-4220-9823-8cc606f5f034","Type":"ContainerStarted","Data":"57a582df620782031dd517ee8549cae414c4157e56c0fc270434b874a8cbfef6"} Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.810454 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-78de-account-create-update-fwjdc" event={"ID":"8b167b56-655a-4220-9823-8cc606f5f034","Type":"ContainerStarted","Data":"c9774d1c3be31829d1732d7995a7c4a40d5cd961beaca133cbe7c41c323af95a"} Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.832257 4948 generic.go:334] "Generic (PLEG): container finished" podID="a5cb8c15-68fc-47db-9f68-1a8401f99f88" containerID="75747717ce5381d273c3b7d9687d0d98efbbf16a7bbea02c0a10888c6dd3a064" exitCode=0 Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.832367 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-2rgc6" event={"ID":"a5cb8c15-68fc-47db-9f68-1a8401f99f88","Type":"ContainerDied","Data":"75747717ce5381d273c3b7d9687d0d98efbbf16a7bbea02c0a10888c6dd3a064"} Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.835481 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.837752 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-7c7cd9876c-262qf" podStartSLOduration=7.837739261 podStartE2EDuration="7.837739261s" podCreationTimestamp="2026-02-20 08:23:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:23:35.822445803 +0000 UTC m=+1064.796940623" watchObservedRunningTime="2026-02-20 08:23:35.837739261 +0000 UTC m=+1064.812234081" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.852639 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-mgnlb" podStartSLOduration=7.85262262 podStartE2EDuration="7.85262262s" podCreationTimestamp="2026-02-20 08:23:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:23:35.85221861 +0000 UTC m=+1064.826713430" watchObservedRunningTime="2026-02-20 08:23:35.85262262 +0000 UTC m=+1064.827117440" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.901923 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-78de-account-create-update-fwjdc" podStartSLOduration=7.901898659 podStartE2EDuration="7.901898659s" podCreationTimestamp="2026-02-20 08:23:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:23:35.871281071 +0000 UTC m=+1064.845775891" watchObservedRunningTime="2026-02-20 08:23:35.901898659 +0000 UTC m=+1064.876393479" Feb 20 08:23:35 crc kubenswrapper[4948]: W0220 08:23:35.924782 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeae40be2_f936_4e13_bc8e_0f84bbcd8edd.slice/crio-8b2a81f0ae61bef074923309a4e18638192928b3bbadabe1634088ddab7ac96c WatchSource:0}: Error finding container 8b2a81f0ae61bef074923309a4e18638192928b3bbadabe1634088ddab7ac96c: Status 404 returned error can't find the container with id 8b2a81f0ae61bef074923309a4e18638192928b3bbadabe1634088ddab7ac96c Feb 20 08:23:36 crc kubenswrapper[4948]: I0220 08:23:36.843276 4948 generic.go:334] "Generic (PLEG): container finished" podID="8b167b56-655a-4220-9823-8cc606f5f034" containerID="57a582df620782031dd517ee8549cae414c4157e56c0fc270434b874a8cbfef6" exitCode=0 Feb 20 08:23:36 crc kubenswrapper[4948]: I0220 08:23:36.843375 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-78de-account-create-update-fwjdc" event={"ID":"8b167b56-655a-4220-9823-8cc606f5f034","Type":"ContainerDied","Data":"57a582df620782031dd517ee8549cae414c4157e56c0fc270434b874a8cbfef6"} Feb 20 08:23:36 crc kubenswrapper[4948]: I0220 08:23:36.845748 4948 generic.go:334] "Generic (PLEG): container finished" podID="b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f" containerID="97e6bc167799c61927aab7b946388571d89bdf96214e87d2ba307142c0a90dd5" exitCode=0 Feb 20 08:23:36 crc kubenswrapper[4948]: I0220 08:23:36.845814 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ca33-account-create-update-dwqpd" event={"ID":"b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f","Type":"ContainerDied","Data":"97e6bc167799c61927aab7b946388571d89bdf96214e87d2ba307142c0a90dd5"} Feb 20 08:23:36 crc kubenswrapper[4948]: I0220 08:23:36.848373 4948 generic.go:334] "Generic (PLEG): container finished" podID="fa398740-9a54-472f-9736-0959dd54b657" containerID="be6725bc5e2da748cc92ca6d3d54cea0585e5ca878e37159db354dc55a1eae54" exitCode=0 Feb 20 08:23:36 crc kubenswrapper[4948]: I0220 08:23:36.848444 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-x2pm7" event={"ID":"fa398740-9a54-472f-9736-0959dd54b657","Type":"ContainerDied","Data":"be6725bc5e2da748cc92ca6d3d54cea0585e5ca878e37159db354dc55a1eae54"} Feb 20 08:23:36 crc kubenswrapper[4948]: I0220 08:23:36.850416 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eae40be2-f936-4e13-bc8e-0f84bbcd8edd","Type":"ContainerStarted","Data":"ef2c6dbff693c9b5a07ebf11c4675349807817c4344b76c4fc763563e5428a3b"} Feb 20 08:23:36 crc kubenswrapper[4948]: I0220 08:23:36.850443 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eae40be2-f936-4e13-bc8e-0f84bbcd8edd","Type":"ContainerStarted","Data":"8b2a81f0ae61bef074923309a4e18638192928b3bbadabe1634088ddab7ac96c"} Feb 20 08:23:36 crc kubenswrapper[4948]: I0220 08:23:36.854419 4948 generic.go:334] "Generic (PLEG): container finished" podID="20b61242-b8c1-4e1b-b319-3eafb79ef048" containerID="ca086fdde2d4c69122e9c700d4d4af3b8218c685639bf6df2d417cc08c1e51d3" exitCode=0 Feb 20 08:23:36 crc kubenswrapper[4948]: I0220 08:23:36.854457 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-mgnlb" event={"ID":"20b61242-b8c1-4e1b-b319-3eafb79ef048","Type":"ContainerDied","Data":"ca086fdde2d4c69122e9c700d4d4af3b8218c685639bf6df2d417cc08c1e51d3"} Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.415446 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-2rgc6" Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.423043 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-aecf-account-create-update-gnnvg" Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.524076 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ttpc\" (UniqueName: \"kubernetes.io/projected/8a311ded-ab56-4f07-8b7f-9dcc4d70f647-kube-api-access-7ttpc\") pod \"8a311ded-ab56-4f07-8b7f-9dcc4d70f647\" (UID: \"8a311ded-ab56-4f07-8b7f-9dcc4d70f647\") " Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.524181 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hgw6\" (UniqueName: \"kubernetes.io/projected/a5cb8c15-68fc-47db-9f68-1a8401f99f88-kube-api-access-2hgw6\") pod \"a5cb8c15-68fc-47db-9f68-1a8401f99f88\" (UID: \"a5cb8c15-68fc-47db-9f68-1a8401f99f88\") " Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.524283 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5cb8c15-68fc-47db-9f68-1a8401f99f88-operator-scripts\") pod \"a5cb8c15-68fc-47db-9f68-1a8401f99f88\" (UID: \"a5cb8c15-68fc-47db-9f68-1a8401f99f88\") " Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.524301 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a311ded-ab56-4f07-8b7f-9dcc4d70f647-operator-scripts\") pod \"8a311ded-ab56-4f07-8b7f-9dcc4d70f647\" (UID: \"8a311ded-ab56-4f07-8b7f-9dcc4d70f647\") " Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.525097 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a311ded-ab56-4f07-8b7f-9dcc4d70f647-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8a311ded-ab56-4f07-8b7f-9dcc4d70f647" (UID: "8a311ded-ab56-4f07-8b7f-9dcc4d70f647"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.525457 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5cb8c15-68fc-47db-9f68-1a8401f99f88-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a5cb8c15-68fc-47db-9f68-1a8401f99f88" (UID: "a5cb8c15-68fc-47db-9f68-1a8401f99f88"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.528747 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a311ded-ab56-4f07-8b7f-9dcc4d70f647-kube-api-access-7ttpc" (OuterVolumeSpecName: "kube-api-access-7ttpc") pod "8a311ded-ab56-4f07-8b7f-9dcc4d70f647" (UID: "8a311ded-ab56-4f07-8b7f-9dcc4d70f647"). InnerVolumeSpecName "kube-api-access-7ttpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.533114 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5cb8c15-68fc-47db-9f68-1a8401f99f88-kube-api-access-2hgw6" (OuterVolumeSpecName: "kube-api-access-2hgw6") pod "a5cb8c15-68fc-47db-9f68-1a8401f99f88" (UID: "a5cb8c15-68fc-47db-9f68-1a8401f99f88"). InnerVolumeSpecName "kube-api-access-2hgw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.626210 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hgw6\" (UniqueName: \"kubernetes.io/projected/a5cb8c15-68fc-47db-9f68-1a8401f99f88-kube-api-access-2hgw6\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.626240 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5cb8c15-68fc-47db-9f68-1a8401f99f88-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.626251 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a311ded-ab56-4f07-8b7f-9dcc4d70f647-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.626262 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ttpc\" (UniqueName: \"kubernetes.io/projected/8a311ded-ab56-4f07-8b7f-9dcc4d70f647-kube-api-access-7ttpc\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.864637 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-2rgc6" event={"ID":"a5cb8c15-68fc-47db-9f68-1a8401f99f88","Type":"ContainerDied","Data":"7de096242d5ba0d916d6903858d689e5e9254091b9036d8255334ff0ea8f1f16"} Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.864661 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-2rgc6" Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.864679 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7de096242d5ba0d916d6903858d689e5e9254091b9036d8255334ff0ea8f1f16" Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.866079 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-aecf-account-create-update-gnnvg" event={"ID":"8a311ded-ab56-4f07-8b7f-9dcc4d70f647","Type":"ContainerDied","Data":"04a106975b06c46c735fa10604e40565f439a58073da7467ca0316db88eec4f9"} Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.866113 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04a106975b06c46c735fa10604e40565f439a58073da7467ca0316db88eec4f9" Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.866095 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-aecf-account-create-update-gnnvg" Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.868158 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eae40be2-f936-4e13-bc8e-0f84bbcd8edd","Type":"ContainerStarted","Data":"74627264af113298ec39e3890cccb5271e59d7d4c945fe853668a4d33ea1db05"} Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.025415 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.025809 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.386601 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-78de-account-create-update-fwjdc" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.535590 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-mgnlb" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.536147 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-x2pm7" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.540325 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpq6v\" (UniqueName: \"kubernetes.io/projected/8b167b56-655a-4220-9823-8cc606f5f034-kube-api-access-fpq6v\") pod \"8b167b56-655a-4220-9823-8cc606f5f034\" (UID: \"8b167b56-655a-4220-9823-8cc606f5f034\") " Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.540420 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b167b56-655a-4220-9823-8cc606f5f034-operator-scripts\") pod \"8b167b56-655a-4220-9823-8cc606f5f034\" (UID: \"8b167b56-655a-4220-9823-8cc606f5f034\") " Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.541169 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b167b56-655a-4220-9823-8cc606f5f034-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8b167b56-655a-4220-9823-8cc606f5f034" (UID: "8b167b56-655a-4220-9823-8cc606f5f034"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.549496 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b167b56-655a-4220-9823-8cc606f5f034-kube-api-access-fpq6v" (OuterVolumeSpecName: "kube-api-access-fpq6v") pod "8b167b56-655a-4220-9823-8cc606f5f034" (UID: "8b167b56-655a-4220-9823-8cc606f5f034"). InnerVolumeSpecName "kube-api-access-fpq6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.550713 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ca33-account-create-update-dwqpd" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.642109 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ssjs\" (UniqueName: \"kubernetes.io/projected/20b61242-b8c1-4e1b-b319-3eafb79ef048-kube-api-access-2ssjs\") pod \"20b61242-b8c1-4e1b-b319-3eafb79ef048\" (UID: \"20b61242-b8c1-4e1b-b319-3eafb79ef048\") " Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.642161 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rx99\" (UniqueName: \"kubernetes.io/projected/fa398740-9a54-472f-9736-0959dd54b657-kube-api-access-7rx99\") pod \"fa398740-9a54-472f-9736-0959dd54b657\" (UID: \"fa398740-9a54-472f-9736-0959dd54b657\") " Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.642249 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vl2t2\" (UniqueName: \"kubernetes.io/projected/b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f-kube-api-access-vl2t2\") pod \"b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f\" (UID: \"b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f\") " Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.642328 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20b61242-b8c1-4e1b-b319-3eafb79ef048-operator-scripts\") pod \"20b61242-b8c1-4e1b-b319-3eafb79ef048\" (UID: \"20b61242-b8c1-4e1b-b319-3eafb79ef048\") " Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.642372 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f-operator-scripts\") pod \"b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f\" (UID: \"b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f\") " Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.642402 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa398740-9a54-472f-9736-0959dd54b657-operator-scripts\") pod \"fa398740-9a54-472f-9736-0959dd54b657\" (UID: \"fa398740-9a54-472f-9736-0959dd54b657\") " Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.643018 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f" (UID: "b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.643445 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa398740-9a54-472f-9736-0959dd54b657-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fa398740-9a54-472f-9736-0959dd54b657" (UID: "fa398740-9a54-472f-9736-0959dd54b657"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.643867 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20b61242-b8c1-4e1b-b319-3eafb79ef048-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "20b61242-b8c1-4e1b-b319-3eafb79ef048" (UID: "20b61242-b8c1-4e1b-b319-3eafb79ef048"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.644201 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpq6v\" (UniqueName: \"kubernetes.io/projected/8b167b56-655a-4220-9823-8cc606f5f034-kube-api-access-fpq6v\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.644216 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20b61242-b8c1-4e1b-b319-3eafb79ef048-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.644227 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.644235 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa398740-9a54-472f-9736-0959dd54b657-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.644243 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b167b56-655a-4220-9823-8cc606f5f034-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.644943 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b61242-b8c1-4e1b-b319-3eafb79ef048-kube-api-access-2ssjs" (OuterVolumeSpecName: "kube-api-access-2ssjs") pod "20b61242-b8c1-4e1b-b319-3eafb79ef048" (UID: "20b61242-b8c1-4e1b-b319-3eafb79ef048"). InnerVolumeSpecName "kube-api-access-2ssjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.647424 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f-kube-api-access-vl2t2" (OuterVolumeSpecName: "kube-api-access-vl2t2") pod "b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f" (UID: "b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f"). InnerVolumeSpecName "kube-api-access-vl2t2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.648449 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa398740-9a54-472f-9736-0959dd54b657-kube-api-access-7rx99" (OuterVolumeSpecName: "kube-api-access-7rx99") pod "fa398740-9a54-472f-9736-0959dd54b657" (UID: "fa398740-9a54-472f-9736-0959dd54b657"). InnerVolumeSpecName "kube-api-access-7rx99". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.746583 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ssjs\" (UniqueName: \"kubernetes.io/projected/20b61242-b8c1-4e1b-b319-3eafb79ef048-kube-api-access-2ssjs\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.747034 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rx99\" (UniqueName: \"kubernetes.io/projected/fa398740-9a54-472f-9736-0959dd54b657-kube-api-access-7rx99\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.747183 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vl2t2\" (UniqueName: \"kubernetes.io/projected/b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f-kube-api-access-vl2t2\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.878876 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-x2pm7" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.878908 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-x2pm7" event={"ID":"fa398740-9a54-472f-9736-0959dd54b657","Type":"ContainerDied","Data":"372eccf775e290867aea35db34cd53f0b147aa6e61ff520ef1ac676e72e0326c"} Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.879302 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="372eccf775e290867aea35db34cd53f0b147aa6e61ff520ef1ac676e72e0326c" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.881696 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eae40be2-f936-4e13-bc8e-0f84bbcd8edd","Type":"ContainerStarted","Data":"e45d96dd2c40e8be3085b92f36d19f6573863121a02ff440da9b19f454a42aef"} Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.883114 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-mgnlb" event={"ID":"20b61242-b8c1-4e1b-b319-3eafb79ef048","Type":"ContainerDied","Data":"c5c2d7e3e5610297feff1dafc766462b7a8ac2739d1c5be80ff63c9dbf1bcc12"} Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.883132 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-mgnlb" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.883143 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c5c2d7e3e5610297feff1dafc766462b7a8ac2739d1c5be80ff63c9dbf1bcc12" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.885059 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-78de-account-create-update-fwjdc" event={"ID":"8b167b56-655a-4220-9823-8cc606f5f034","Type":"ContainerDied","Data":"c9774d1c3be31829d1732d7995a7c4a40d5cd961beaca133cbe7c41c323af95a"} Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.885102 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9774d1c3be31829d1732d7995a7c4a40d5cd961beaca133cbe7c41c323af95a" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.885073 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-78de-account-create-update-fwjdc" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.887018 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ca33-account-create-update-dwqpd" event={"ID":"b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f","Type":"ContainerDied","Data":"4a6359d724bfd5ae6bb40a902ffdabe96c01743e8c81b47b0e26ad235b7eaeff"} Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.887052 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a6359d724bfd5ae6bb40a902ffdabe96c01743e8c81b47b0e26ad235b7eaeff" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.887092 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ca33-account-create-update-dwqpd" Feb 20 08:23:39 crc kubenswrapper[4948]: I0220 08:23:39.041429 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:40 crc kubenswrapper[4948]: I0220 08:23:40.905358 4948 generic.go:334] "Generic (PLEG): container finished" podID="d1046503-c6d3-49bf-9523-2f897fcb82ba" containerID="577865ca69f0d8457d617948e5cac2be0a124e4620edc60df1395130de8a2334" exitCode=137 Feb 20 08:23:40 crc kubenswrapper[4948]: I0220 08:23:40.905535 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66b4d4cb8-hgmkd" event={"ID":"d1046503-c6d3-49bf-9523-2f897fcb82ba","Type":"ContainerDied","Data":"577865ca69f0d8457d617948e5cac2be0a124e4620edc60df1395130de8a2334"} Feb 20 08:23:40 crc kubenswrapper[4948]: I0220 08:23:40.906091 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66b4d4cb8-hgmkd" event={"ID":"d1046503-c6d3-49bf-9523-2f897fcb82ba","Type":"ContainerDied","Data":"0fc9151c9d2a99b81f2c9346beebf0499347807bacbd7c95da55982c7422c38e"} Feb 20 08:23:40 crc kubenswrapper[4948]: I0220 08:23:40.906110 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0fc9151c9d2a99b81f2c9346beebf0499347807bacbd7c95da55982c7422c38e" Feb 20 08:23:40 crc kubenswrapper[4948]: I0220 08:23:40.909132 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eae40be2-f936-4e13-bc8e-0f84bbcd8edd","Type":"ContainerStarted","Data":"52810114bb238a5ef2002a9ed77afb4259619263e919c688def5a10a52ac2cbd"} Feb 20 08:23:40 crc kubenswrapper[4948]: I0220 08:23:40.909284 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerName="ceilometer-central-agent" containerID="cri-o://ef2c6dbff693c9b5a07ebf11c4675349807817c4344b76c4fc763563e5428a3b" gracePeriod=30 Feb 20 08:23:40 crc kubenswrapper[4948]: I0220 08:23:40.909373 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 20 08:23:40 crc kubenswrapper[4948]: I0220 08:23:40.909693 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerName="proxy-httpd" containerID="cri-o://52810114bb238a5ef2002a9ed77afb4259619263e919c688def5a10a52ac2cbd" gracePeriod=30 Feb 20 08:23:40 crc kubenswrapper[4948]: I0220 08:23:40.909743 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerName="sg-core" containerID="cri-o://e45d96dd2c40e8be3085b92f36d19f6573863121a02ff440da9b19f454a42aef" gracePeriod=30 Feb 20 08:23:40 crc kubenswrapper[4948]: I0220 08:23:40.909780 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerName="ceilometer-notification-agent" containerID="cri-o://74627264af113298ec39e3890cccb5271e59d7d4c945fe853668a4d33ea1db05" gracePeriod=30 Feb 20 08:23:40 crc kubenswrapper[4948]: I0220 08:23:40.937271 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.963706826 podStartE2EDuration="6.937251168s" podCreationTimestamp="2026-02-20 08:23:34 +0000 UTC" firstStartedPulling="2026-02-20 08:23:35.92817329 +0000 UTC m=+1064.902668110" lastFinishedPulling="2026-02-20 08:23:39.901717632 +0000 UTC m=+1068.876212452" observedRunningTime="2026-02-20 08:23:40.937225837 +0000 UTC m=+1069.911720657" watchObservedRunningTime="2026-02-20 08:23:40.937251168 +0000 UTC m=+1069.911745988" Feb 20 08:23:40 crc kubenswrapper[4948]: I0220 08:23:40.978793 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.097170 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-horizon-secret-key\") pod \"d1046503-c6d3-49bf-9523-2f897fcb82ba\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.097237 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgwfv\" (UniqueName: \"kubernetes.io/projected/d1046503-c6d3-49bf-9523-2f897fcb82ba-kube-api-access-mgwfv\") pod \"d1046503-c6d3-49bf-9523-2f897fcb82ba\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.097286 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-horizon-tls-certs\") pod \"d1046503-c6d3-49bf-9523-2f897fcb82ba\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.097432 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d1046503-c6d3-49bf-9523-2f897fcb82ba-logs\") pod \"d1046503-c6d3-49bf-9523-2f897fcb82ba\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.097456 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-combined-ca-bundle\") pod \"d1046503-c6d3-49bf-9523-2f897fcb82ba\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.097478 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d1046503-c6d3-49bf-9523-2f897fcb82ba-config-data\") pod \"d1046503-c6d3-49bf-9523-2f897fcb82ba\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.097541 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d1046503-c6d3-49bf-9523-2f897fcb82ba-scripts\") pod \"d1046503-c6d3-49bf-9523-2f897fcb82ba\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.098043 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1046503-c6d3-49bf-9523-2f897fcb82ba-logs" (OuterVolumeSpecName: "logs") pod "d1046503-c6d3-49bf-9523-2f897fcb82ba" (UID: "d1046503-c6d3-49bf-9523-2f897fcb82ba"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.103148 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "d1046503-c6d3-49bf-9523-2f897fcb82ba" (UID: "d1046503-c6d3-49bf-9523-2f897fcb82ba"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.103957 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1046503-c6d3-49bf-9523-2f897fcb82ba-kube-api-access-mgwfv" (OuterVolumeSpecName: "kube-api-access-mgwfv") pod "d1046503-c6d3-49bf-9523-2f897fcb82ba" (UID: "d1046503-c6d3-49bf-9523-2f897fcb82ba"). InnerVolumeSpecName "kube-api-access-mgwfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.124021 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1046503-c6d3-49bf-9523-2f897fcb82ba-scripts" (OuterVolumeSpecName: "scripts") pod "d1046503-c6d3-49bf-9523-2f897fcb82ba" (UID: "d1046503-c6d3-49bf-9523-2f897fcb82ba"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.127777 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d1046503-c6d3-49bf-9523-2f897fcb82ba" (UID: "d1046503-c6d3-49bf-9523-2f897fcb82ba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.128788 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1046503-c6d3-49bf-9523-2f897fcb82ba-config-data" (OuterVolumeSpecName: "config-data") pod "d1046503-c6d3-49bf-9523-2f897fcb82ba" (UID: "d1046503-c6d3-49bf-9523-2f897fcb82ba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.161126 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "d1046503-c6d3-49bf-9523-2f897fcb82ba" (UID: "d1046503-c6d3-49bf-9523-2f897fcb82ba"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.199318 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgwfv\" (UniqueName: \"kubernetes.io/projected/d1046503-c6d3-49bf-9523-2f897fcb82ba-kube-api-access-mgwfv\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.199359 4948 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.199368 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d1046503-c6d3-49bf-9523-2f897fcb82ba-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.199377 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.199386 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d1046503-c6d3-49bf-9523-2f897fcb82ba-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.199394 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d1046503-c6d3-49bf-9523-2f897fcb82ba-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.199402 4948 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.921388 4948 generic.go:334] "Generic (PLEG): container finished" podID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerID="52810114bb238a5ef2002a9ed77afb4259619263e919c688def5a10a52ac2cbd" exitCode=0 Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.921752 4948 generic.go:334] "Generic (PLEG): container finished" podID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerID="e45d96dd2c40e8be3085b92f36d19f6573863121a02ff440da9b19f454a42aef" exitCode=2 Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.921768 4948 generic.go:334] "Generic (PLEG): container finished" podID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerID="74627264af113298ec39e3890cccb5271e59d7d4c945fe853668a4d33ea1db05" exitCode=0 Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.921554 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eae40be2-f936-4e13-bc8e-0f84bbcd8edd","Type":"ContainerDied","Data":"52810114bb238a5ef2002a9ed77afb4259619263e919c688def5a10a52ac2cbd"} Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.921849 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.921895 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eae40be2-f936-4e13-bc8e-0f84bbcd8edd","Type":"ContainerDied","Data":"e45d96dd2c40e8be3085b92f36d19f6573863121a02ff440da9b19f454a42aef"} Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.921925 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eae40be2-f936-4e13-bc8e-0f84bbcd8edd","Type":"ContainerDied","Data":"74627264af113298ec39e3890cccb5271e59d7d4c945fe853668a4d33ea1db05"} Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.961507 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-66b4d4cb8-hgmkd"] Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.969336 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-66b4d4cb8-hgmkd"] Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.369549 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.527376 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-config-data\") pod \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.527608 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-scripts\") pod \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.527712 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qc4df\" (UniqueName: \"kubernetes.io/projected/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-kube-api-access-qc4df\") pod \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.527767 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-run-httpd\") pod \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.527853 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-log-httpd\") pod \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.527926 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-sg-core-conf-yaml\") pod \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.527996 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-combined-ca-bundle\") pod \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.528859 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "eae40be2-f936-4e13-bc8e-0f84bbcd8edd" (UID: "eae40be2-f936-4e13-bc8e-0f84bbcd8edd"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.530683 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "eae40be2-f936-4e13-bc8e-0f84bbcd8edd" (UID: "eae40be2-f936-4e13-bc8e-0f84bbcd8edd"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.538964 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-kube-api-access-qc4df" (OuterVolumeSpecName: "kube-api-access-qc4df") pod "eae40be2-f936-4e13-bc8e-0f84bbcd8edd" (UID: "eae40be2-f936-4e13-bc8e-0f84bbcd8edd"). InnerVolumeSpecName "kube-api-access-qc4df". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.539450 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-scripts" (OuterVolumeSpecName: "scripts") pod "eae40be2-f936-4e13-bc8e-0f84bbcd8edd" (UID: "eae40be2-f936-4e13-bc8e-0f84bbcd8edd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.559358 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "eae40be2-f936-4e13-bc8e-0f84bbcd8edd" (UID: "eae40be2-f936-4e13-bc8e-0f84bbcd8edd"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.624332 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eae40be2-f936-4e13-bc8e-0f84bbcd8edd" (UID: "eae40be2-f936-4e13-bc8e-0f84bbcd8edd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.630034 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.630064 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qc4df\" (UniqueName: \"kubernetes.io/projected/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-kube-api-access-qc4df\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.630074 4948 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.630083 4948 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.630092 4948 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.630101 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.666200 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-config-data" (OuterVolumeSpecName: "config-data") pod "eae40be2-f936-4e13-bc8e-0f84bbcd8edd" (UID: "eae40be2-f936-4e13-bc8e-0f84bbcd8edd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.731339 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.756517 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.813994 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-b5c5cc45d-7dcbq"] Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.814270 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-b5c5cc45d-7dcbq" podUID="fd1b85e2-1b5f-4993-bc66-72daaf6aa79d" containerName="neutron-api" containerID="cri-o://fceb35ea8befbc40c2edb25ea5ef78a9c4a759e83b13e451735a1e2ccf6a99a8" gracePeriod=30 Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.814739 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-b5c5cc45d-7dcbq" podUID="fd1b85e2-1b5f-4993-bc66-72daaf6aa79d" containerName="neutron-httpd" containerID="cri-o://4fc7cff150a17fca1672ea7d9eae41c85a175b0d1f41224639307a4d7025cd12" gracePeriod=30 Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.935510 4948 generic.go:334] "Generic (PLEG): container finished" podID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerID="ef2c6dbff693c9b5a07ebf11c4675349807817c4344b76c4fc763563e5428a3b" exitCode=0 Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.935546 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eae40be2-f936-4e13-bc8e-0f84bbcd8edd","Type":"ContainerDied","Data":"ef2c6dbff693c9b5a07ebf11c4675349807817c4344b76c4fc763563e5428a3b"} Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.935570 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eae40be2-f936-4e13-bc8e-0f84bbcd8edd","Type":"ContainerDied","Data":"8b2a81f0ae61bef074923309a4e18638192928b3bbadabe1634088ddab7ac96c"} Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.935586 4948 scope.go:117] "RemoveContainer" containerID="52810114bb238a5ef2002a9ed77afb4259619263e919c688def5a10a52ac2cbd" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.935705 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.966242 4948 scope.go:117] "RemoveContainer" containerID="e45d96dd2c40e8be3085b92f36d19f6573863121a02ff440da9b19f454a42aef" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.969097 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.976769 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.004307 4948 scope.go:117] "RemoveContainer" containerID="74627264af113298ec39e3890cccb5271e59d7d4c945fe853668a4d33ea1db05" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.004422 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:43 crc kubenswrapper[4948]: E0220 08:23:43.004785 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f" containerName="mariadb-account-create-update" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.004800 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f" containerName="mariadb-account-create-update" Feb 20 08:23:43 crc kubenswrapper[4948]: E0220 08:23:43.004811 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1046503-c6d3-49bf-9523-2f897fcb82ba" containerName="horizon-log" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.004817 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1046503-c6d3-49bf-9523-2f897fcb82ba" containerName="horizon-log" Feb 20 08:23:43 crc kubenswrapper[4948]: E0220 08:23:43.004826 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b167b56-655a-4220-9823-8cc606f5f034" containerName="mariadb-account-create-update" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.004833 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b167b56-655a-4220-9823-8cc606f5f034" containerName="mariadb-account-create-update" Feb 20 08:23:43 crc kubenswrapper[4948]: E0220 08:23:43.004841 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerName="ceilometer-central-agent" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.004847 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerName="ceilometer-central-agent" Feb 20 08:23:43 crc kubenswrapper[4948]: E0220 08:23:43.004856 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerName="ceilometer-notification-agent" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.004863 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerName="ceilometer-notification-agent" Feb 20 08:23:43 crc kubenswrapper[4948]: E0220 08:23:43.004877 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a311ded-ab56-4f07-8b7f-9dcc4d70f647" containerName="mariadb-account-create-update" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.004883 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a311ded-ab56-4f07-8b7f-9dcc4d70f647" containerName="mariadb-account-create-update" Feb 20 08:23:43 crc kubenswrapper[4948]: E0220 08:23:43.004899 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerName="proxy-httpd" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.004905 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerName="proxy-httpd" Feb 20 08:23:43 crc kubenswrapper[4948]: E0220 08:23:43.004921 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5cb8c15-68fc-47db-9f68-1a8401f99f88" containerName="mariadb-database-create" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.004927 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5cb8c15-68fc-47db-9f68-1a8401f99f88" containerName="mariadb-database-create" Feb 20 08:23:43 crc kubenswrapper[4948]: E0220 08:23:43.004935 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerName="sg-core" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.004942 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerName="sg-core" Feb 20 08:23:43 crc kubenswrapper[4948]: E0220 08:23:43.004950 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20b61242-b8c1-4e1b-b319-3eafb79ef048" containerName="mariadb-database-create" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.004956 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="20b61242-b8c1-4e1b-b319-3eafb79ef048" containerName="mariadb-database-create" Feb 20 08:23:43 crc kubenswrapper[4948]: E0220 08:23:43.004964 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1046503-c6d3-49bf-9523-2f897fcb82ba" containerName="horizon" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.005016 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1046503-c6d3-49bf-9523-2f897fcb82ba" containerName="horizon" Feb 20 08:23:43 crc kubenswrapper[4948]: E0220 08:23:43.005028 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa398740-9a54-472f-9736-0959dd54b657" containerName="mariadb-database-create" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.005035 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa398740-9a54-472f-9736-0959dd54b657" containerName="mariadb-database-create" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.006376 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerName="ceilometer-notification-agent" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.006400 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1046503-c6d3-49bf-9523-2f897fcb82ba" containerName="horizon-log" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.006413 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="20b61242-b8c1-4e1b-b319-3eafb79ef048" containerName="mariadb-database-create" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.006423 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b167b56-655a-4220-9823-8cc606f5f034" containerName="mariadb-account-create-update" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.006445 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a311ded-ab56-4f07-8b7f-9dcc4d70f647" containerName="mariadb-account-create-update" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.006457 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerName="proxy-httpd" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.006470 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f" containerName="mariadb-account-create-update" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.006483 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5cb8c15-68fc-47db-9f68-1a8401f99f88" containerName="mariadb-database-create" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.006498 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerName="ceilometer-central-agent" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.006507 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa398740-9a54-472f-9736-0959dd54b657" containerName="mariadb-database-create" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.006521 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerName="sg-core" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.006532 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1046503-c6d3-49bf-9523-2f897fcb82ba" containerName="horizon" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.008566 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.012347 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.012572 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.033769 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.040108 4948 scope.go:117] "RemoveContainer" containerID="ef2c6dbff693c9b5a07ebf11c4675349807817c4344b76c4fc763563e5428a3b" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.096330 4948 scope.go:117] "RemoveContainer" containerID="52810114bb238a5ef2002a9ed77afb4259619263e919c688def5a10a52ac2cbd" Feb 20 08:23:43 crc kubenswrapper[4948]: E0220 08:23:43.096684 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52810114bb238a5ef2002a9ed77afb4259619263e919c688def5a10a52ac2cbd\": container with ID starting with 52810114bb238a5ef2002a9ed77afb4259619263e919c688def5a10a52ac2cbd not found: ID does not exist" containerID="52810114bb238a5ef2002a9ed77afb4259619263e919c688def5a10a52ac2cbd" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.096718 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52810114bb238a5ef2002a9ed77afb4259619263e919c688def5a10a52ac2cbd"} err="failed to get container status \"52810114bb238a5ef2002a9ed77afb4259619263e919c688def5a10a52ac2cbd\": rpc error: code = NotFound desc = could not find container \"52810114bb238a5ef2002a9ed77afb4259619263e919c688def5a10a52ac2cbd\": container with ID starting with 52810114bb238a5ef2002a9ed77afb4259619263e919c688def5a10a52ac2cbd not found: ID does not exist" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.096738 4948 scope.go:117] "RemoveContainer" containerID="e45d96dd2c40e8be3085b92f36d19f6573863121a02ff440da9b19f454a42aef" Feb 20 08:23:43 crc kubenswrapper[4948]: E0220 08:23:43.096940 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e45d96dd2c40e8be3085b92f36d19f6573863121a02ff440da9b19f454a42aef\": container with ID starting with e45d96dd2c40e8be3085b92f36d19f6573863121a02ff440da9b19f454a42aef not found: ID does not exist" containerID="e45d96dd2c40e8be3085b92f36d19f6573863121a02ff440da9b19f454a42aef" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.096963 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e45d96dd2c40e8be3085b92f36d19f6573863121a02ff440da9b19f454a42aef"} err="failed to get container status \"e45d96dd2c40e8be3085b92f36d19f6573863121a02ff440da9b19f454a42aef\": rpc error: code = NotFound desc = could not find container \"e45d96dd2c40e8be3085b92f36d19f6573863121a02ff440da9b19f454a42aef\": container with ID starting with e45d96dd2c40e8be3085b92f36d19f6573863121a02ff440da9b19f454a42aef not found: ID does not exist" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.096990 4948 scope.go:117] "RemoveContainer" containerID="74627264af113298ec39e3890cccb5271e59d7d4c945fe853668a4d33ea1db05" Feb 20 08:23:43 crc kubenswrapper[4948]: E0220 08:23:43.097283 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74627264af113298ec39e3890cccb5271e59d7d4c945fe853668a4d33ea1db05\": container with ID starting with 74627264af113298ec39e3890cccb5271e59d7d4c945fe853668a4d33ea1db05 not found: ID does not exist" containerID="74627264af113298ec39e3890cccb5271e59d7d4c945fe853668a4d33ea1db05" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.097321 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74627264af113298ec39e3890cccb5271e59d7d4c945fe853668a4d33ea1db05"} err="failed to get container status \"74627264af113298ec39e3890cccb5271e59d7d4c945fe853668a4d33ea1db05\": rpc error: code = NotFound desc = could not find container \"74627264af113298ec39e3890cccb5271e59d7d4c945fe853668a4d33ea1db05\": container with ID starting with 74627264af113298ec39e3890cccb5271e59d7d4c945fe853668a4d33ea1db05 not found: ID does not exist" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.097345 4948 scope.go:117] "RemoveContainer" containerID="ef2c6dbff693c9b5a07ebf11c4675349807817c4344b76c4fc763563e5428a3b" Feb 20 08:23:43 crc kubenswrapper[4948]: E0220 08:23:43.097551 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef2c6dbff693c9b5a07ebf11c4675349807817c4344b76c4fc763563e5428a3b\": container with ID starting with ef2c6dbff693c9b5a07ebf11c4675349807817c4344b76c4fc763563e5428a3b not found: ID does not exist" containerID="ef2c6dbff693c9b5a07ebf11c4675349807817c4344b76c4fc763563e5428a3b" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.097571 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef2c6dbff693c9b5a07ebf11c4675349807817c4344b76c4fc763563e5428a3b"} err="failed to get container status \"ef2c6dbff693c9b5a07ebf11c4675349807817c4344b76c4fc763563e5428a3b\": rpc error: code = NotFound desc = could not find container \"ef2c6dbff693c9b5a07ebf11c4675349807817c4344b76c4fc763563e5428a3b\": container with ID starting with ef2c6dbff693c9b5a07ebf11c4675349807817c4344b76c4fc763563e5428a3b not found: ID does not exist" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.149118 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-config-data\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.149182 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txt8k\" (UniqueName: \"kubernetes.io/projected/0894350a-0df8-4c04-b7c7-fd8f93139959-kube-api-access-txt8k\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.149206 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0894350a-0df8-4c04-b7c7-fd8f93139959-log-httpd\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.149386 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.149444 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0894350a-0df8-4c04-b7c7-fd8f93139959-run-httpd\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.149604 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-scripts\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.149687 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.251606 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-scripts\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.251666 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.251705 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-config-data\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.251747 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txt8k\" (UniqueName: \"kubernetes.io/projected/0894350a-0df8-4c04-b7c7-fd8f93139959-kube-api-access-txt8k\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.251775 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0894350a-0df8-4c04-b7c7-fd8f93139959-log-httpd\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.251814 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.251834 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0894350a-0df8-4c04-b7c7-fd8f93139959-run-httpd\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.252308 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0894350a-0df8-4c04-b7c7-fd8f93139959-run-httpd\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.252441 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0894350a-0df8-4c04-b7c7-fd8f93139959-log-httpd\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.256737 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-scripts\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.257777 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-config-data\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.258517 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.259321 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.277342 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txt8k\" (UniqueName: \"kubernetes.io/projected/0894350a-0df8-4c04-b7c7-fd8f93139959-kube-api-access-txt8k\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.332020 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.736277 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1046503-c6d3-49bf-9523-2f897fcb82ba" path="/var/lib/kubelet/pods/d1046503-c6d3-49bf-9523-2f897fcb82ba/volumes" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.737226 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" path="/var/lib/kubelet/pods/eae40be2-f936-4e13-bc8e-0f84bbcd8edd/volumes" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.813709 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:43 crc kubenswrapper[4948]: W0220 08:23:43.818395 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0894350a_0df8_4c04_b7c7_fd8f93139959.slice/crio-6657eaf082a861ad8f0f29c4c00aaf35d258d94004b82734a54c5722b14133f5 WatchSource:0}: Error finding container 6657eaf082a861ad8f0f29c4c00aaf35d258d94004b82734a54c5722b14133f5: Status 404 returned error can't find the container with id 6657eaf082a861ad8f0f29c4c00aaf35d258d94004b82734a54c5722b14133f5 Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.946620 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0894350a-0df8-4c04-b7c7-fd8f93139959","Type":"ContainerStarted","Data":"6657eaf082a861ad8f0f29c4c00aaf35d258d94004b82734a54c5722b14133f5"} Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.952506 4948 generic.go:334] "Generic (PLEG): container finished" podID="fd1b85e2-1b5f-4993-bc66-72daaf6aa79d" containerID="4fc7cff150a17fca1672ea7d9eae41c85a175b0d1f41224639307a4d7025cd12" exitCode=0 Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.952590 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b5c5cc45d-7dcbq" event={"ID":"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d","Type":"ContainerDied","Data":"4fc7cff150a17fca1672ea7d9eae41c85a175b0d1f41224639307a4d7025cd12"} Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.968134 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.973728 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.518730 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xt5t7"] Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.519859 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xt5t7" Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.523657 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.523785 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-2tb2s" Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.523933 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.573606 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xt5t7"] Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.677393 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-config-data\") pod \"nova-cell0-conductor-db-sync-xt5t7\" (UID: \"3fdb2e02-8713-4205-ae31-0af33f8e4047\") " pod="openstack/nova-cell0-conductor-db-sync-xt5t7" Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.677516 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4jh4\" (UniqueName: \"kubernetes.io/projected/3fdb2e02-8713-4205-ae31-0af33f8e4047-kube-api-access-j4jh4\") pod \"nova-cell0-conductor-db-sync-xt5t7\" (UID: \"3fdb2e02-8713-4205-ae31-0af33f8e4047\") " pod="openstack/nova-cell0-conductor-db-sync-xt5t7" Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.677542 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-xt5t7\" (UID: \"3fdb2e02-8713-4205-ae31-0af33f8e4047\") " pod="openstack/nova-cell0-conductor-db-sync-xt5t7" Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.677559 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-scripts\") pod \"nova-cell0-conductor-db-sync-xt5t7\" (UID: \"3fdb2e02-8713-4205-ae31-0af33f8e4047\") " pod="openstack/nova-cell0-conductor-db-sync-xt5t7" Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.780116 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-config-data\") pod \"nova-cell0-conductor-db-sync-xt5t7\" (UID: \"3fdb2e02-8713-4205-ae31-0af33f8e4047\") " pod="openstack/nova-cell0-conductor-db-sync-xt5t7" Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.781401 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4jh4\" (UniqueName: \"kubernetes.io/projected/3fdb2e02-8713-4205-ae31-0af33f8e4047-kube-api-access-j4jh4\") pod \"nova-cell0-conductor-db-sync-xt5t7\" (UID: \"3fdb2e02-8713-4205-ae31-0af33f8e4047\") " pod="openstack/nova-cell0-conductor-db-sync-xt5t7" Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.781450 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-xt5t7\" (UID: \"3fdb2e02-8713-4205-ae31-0af33f8e4047\") " pod="openstack/nova-cell0-conductor-db-sync-xt5t7" Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.781483 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-scripts\") pod \"nova-cell0-conductor-db-sync-xt5t7\" (UID: \"3fdb2e02-8713-4205-ae31-0af33f8e4047\") " pod="openstack/nova-cell0-conductor-db-sync-xt5t7" Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.786813 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-scripts\") pod \"nova-cell0-conductor-db-sync-xt5t7\" (UID: \"3fdb2e02-8713-4205-ae31-0af33f8e4047\") " pod="openstack/nova-cell0-conductor-db-sync-xt5t7" Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.786938 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-config-data\") pod \"nova-cell0-conductor-db-sync-xt5t7\" (UID: \"3fdb2e02-8713-4205-ae31-0af33f8e4047\") " pod="openstack/nova-cell0-conductor-db-sync-xt5t7" Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.788651 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-xt5t7\" (UID: \"3fdb2e02-8713-4205-ae31-0af33f8e4047\") " pod="openstack/nova-cell0-conductor-db-sync-xt5t7" Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.800822 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4jh4\" (UniqueName: \"kubernetes.io/projected/3fdb2e02-8713-4205-ae31-0af33f8e4047-kube-api-access-j4jh4\") pod \"nova-cell0-conductor-db-sync-xt5t7\" (UID: \"3fdb2e02-8713-4205-ae31-0af33f8e4047\") " pod="openstack/nova-cell0-conductor-db-sync-xt5t7" Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.844755 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xt5t7" Feb 20 08:23:45 crc kubenswrapper[4948]: I0220 08:23:45.302892 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xt5t7"] Feb 20 08:23:45 crc kubenswrapper[4948]: I0220 08:23:45.984648 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xt5t7" event={"ID":"3fdb2e02-8713-4205-ae31-0af33f8e4047","Type":"ContainerStarted","Data":"d20f6a9cd2d1756d07c023cf4fc3e418531e5c20161aa674031ec3aa6a8df0d2"} Feb 20 08:23:46 crc kubenswrapper[4948]: I0220 08:23:46.911850 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:46 crc kubenswrapper[4948]: I0220 08:23:46.995248 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0894350a-0df8-4c04-b7c7-fd8f93139959","Type":"ContainerStarted","Data":"8372790439c71a0ea7e4da810ca880dff1b44a6d3fa022b0a9cbbafbf8e763cc"} Feb 20 08:23:46 crc kubenswrapper[4948]: I0220 08:23:46.995286 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0894350a-0df8-4c04-b7c7-fd8f93139959","Type":"ContainerStarted","Data":"5a880cf140d0830077bc8b3d8a9a05a1bd68a08c91042ee55e74a3422f78c6e4"} Feb 20 08:23:48 crc kubenswrapper[4948]: I0220 08:23:48.016249 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0894350a-0df8-4c04-b7c7-fd8f93139959","Type":"ContainerStarted","Data":"d2ec8efadba57da267b73b627d540acadced09eb54d33e0b4318c801f4b817ed"} Feb 20 08:23:49 crc kubenswrapper[4948]: I0220 08:23:49.027062 4948 generic.go:334] "Generic (PLEG): container finished" podID="fd1b85e2-1b5f-4993-bc66-72daaf6aa79d" containerID="fceb35ea8befbc40c2edb25ea5ef78a9c4a759e83b13e451735a1e2ccf6a99a8" exitCode=0 Feb 20 08:23:49 crc kubenswrapper[4948]: I0220 08:23:49.027148 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b5c5cc45d-7dcbq" event={"ID":"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d","Type":"ContainerDied","Data":"fceb35ea8befbc40c2edb25ea5ef78a9c4a759e83b13e451735a1e2ccf6a99a8"} Feb 20 08:23:51 crc kubenswrapper[4948]: I0220 08:23:51.350279 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:51 crc kubenswrapper[4948]: I0220 08:23:51.365393 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:51 crc kubenswrapper[4948]: I0220 08:23:51.427321 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-d46988dcb-rn6qs"] Feb 20 08:23:51 crc kubenswrapper[4948]: I0220 08:23:51.427864 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-d46988dcb-rn6qs" podUID="8b1ee525-1a6e-461d-aa4d-178ab601ec9d" containerName="placement-log" containerID="cri-o://26cbea34a0a499d9db14acd1bf75d24dc0d0a677df528028730bacad671ebd26" gracePeriod=30 Feb 20 08:23:51 crc kubenswrapper[4948]: I0220 08:23:51.428021 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-d46988dcb-rn6qs" podUID="8b1ee525-1a6e-461d-aa4d-178ab601ec9d" containerName="placement-api" containerID="cri-o://6a2830020d67b39ba1edbe811de2ebbd05b2b8a02d0578d95553d45d985544af" gracePeriod=30 Feb 20 08:23:52 crc kubenswrapper[4948]: I0220 08:23:52.072119 4948 generic.go:334] "Generic (PLEG): container finished" podID="8b1ee525-1a6e-461d-aa4d-178ab601ec9d" containerID="26cbea34a0a499d9db14acd1bf75d24dc0d0a677df528028730bacad671ebd26" exitCode=143 Feb 20 08:23:52 crc kubenswrapper[4948]: I0220 08:23:52.072205 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d46988dcb-rn6qs" event={"ID":"8b1ee525-1a6e-461d-aa4d-178ab601ec9d","Type":"ContainerDied","Data":"26cbea34a0a499d9db14acd1bf75d24dc0d0a677df528028730bacad671ebd26"} Feb 20 08:23:52 crc kubenswrapper[4948]: I0220 08:23:52.406201 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:23:52 crc kubenswrapper[4948]: I0220 08:23:52.406428 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d9399b9c-0935-4735-8b08-96d4a29d4ba8" containerName="glance-log" containerID="cri-o://a66ddb00a7f57dc492a3a8a03a3bce06f45c65381b1e5db4f923e10fec5267a0" gracePeriod=30 Feb 20 08:23:52 crc kubenswrapper[4948]: I0220 08:23:52.406545 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d9399b9c-0935-4735-8b08-96d4a29d4ba8" containerName="glance-httpd" containerID="cri-o://656e84496ca041e8a454b07ed67e165b130af1030016564ae4519b8e2e6f11e5" gracePeriod=30 Feb 20 08:23:53 crc kubenswrapper[4948]: I0220 08:23:53.084216 4948 generic.go:334] "Generic (PLEG): container finished" podID="d9399b9c-0935-4735-8b08-96d4a29d4ba8" containerID="a66ddb00a7f57dc492a3a8a03a3bce06f45c65381b1e5db4f923e10fec5267a0" exitCode=143 Feb 20 08:23:53 crc kubenswrapper[4948]: I0220 08:23:53.084249 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d9399b9c-0935-4735-8b08-96d4a29d4ba8","Type":"ContainerDied","Data":"a66ddb00a7f57dc492a3a8a03a3bce06f45c65381b1e5db4f923e10fec5267a0"} Feb 20 08:23:53 crc kubenswrapper[4948]: I0220 08:23:53.225834 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 08:23:53 crc kubenswrapper[4948]: I0220 08:23:53.226161 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="33b35605-1057-4e64-99a7-a7273351d6c2" containerName="glance-log" containerID="cri-o://9232d9b3b9f362e1b01dbc32c5300081164bdbed5e40b2c4cd59a5e4f4b28166" gracePeriod=30 Feb 20 08:23:53 crc kubenswrapper[4948]: I0220 08:23:53.226626 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="33b35605-1057-4e64-99a7-a7273351d6c2" containerName="glance-httpd" containerID="cri-o://08ec8e2a4374c5a50916bcc6e4fbc0d7957aac63564273dd75ed243c1b2a0679" gracePeriod=30 Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.107849 4948 generic.go:334] "Generic (PLEG): container finished" podID="33b35605-1057-4e64-99a7-a7273351d6c2" containerID="9232d9b3b9f362e1b01dbc32c5300081164bdbed5e40b2c4cd59a5e4f4b28166" exitCode=143 Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.107948 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"33b35605-1057-4e64-99a7-a7273351d6c2","Type":"ContainerDied","Data":"9232d9b3b9f362e1b01dbc32c5300081164bdbed5e40b2c4cd59a5e4f4b28166"} Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.113846 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerName="proxy-httpd" containerID="cri-o://c4e7730b1940fa1741db63557d39a1291cbd6edeee74b043d33b1623adece0d5" gracePeriod=30 Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.113877 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.113843 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerName="ceilometer-central-agent" containerID="cri-o://5a880cf140d0830077bc8b3d8a9a05a1bd68a08c91042ee55e74a3422f78c6e4" gracePeriod=30 Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.113936 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerName="sg-core" containerID="cri-o://d2ec8efadba57da267b73b627d540acadced09eb54d33e0b4318c801f4b817ed" gracePeriod=30 Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.113943 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerName="ceilometer-notification-agent" containerID="cri-o://8372790439c71a0ea7e4da810ca880dff1b44a6d3fa022b0a9cbbafbf8e763cc" gracePeriod=30 Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.114195 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.115648 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xt5t7" event={"ID":"3fdb2e02-8713-4205-ae31-0af33f8e4047","Type":"ContainerStarted","Data":"53a1cca1717eaf7448d7d84b4b68c8bad0f23b2d1cc07d696f16d15b417b8356"} Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.152638 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.105361799 podStartE2EDuration="12.152620418s" podCreationTimestamp="2026-02-20 08:23:42 +0000 UTC" firstStartedPulling="2026-02-20 08:23:43.821928265 +0000 UTC m=+1072.796423085" lastFinishedPulling="2026-02-20 08:23:53.869186884 +0000 UTC m=+1082.843681704" observedRunningTime="2026-02-20 08:23:54.144107657 +0000 UTC m=+1083.118602477" watchObservedRunningTime="2026-02-20 08:23:54.152620418 +0000 UTC m=+1083.127115238" Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.157730 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-ovndb-tls-certs\") pod \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.157846 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-combined-ca-bundle\") pod \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.157935 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khls2\" (UniqueName: \"kubernetes.io/projected/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-kube-api-access-khls2\") pod \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.157989 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-httpd-config\") pod \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.158137 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-config\") pod \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.166238 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "fd1b85e2-1b5f-4993-bc66-72daaf6aa79d" (UID: "fd1b85e2-1b5f-4993-bc66-72daaf6aa79d"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.166487 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-kube-api-access-khls2" (OuterVolumeSpecName: "kube-api-access-khls2") pod "fd1b85e2-1b5f-4993-bc66-72daaf6aa79d" (UID: "fd1b85e2-1b5f-4993-bc66-72daaf6aa79d"). InnerVolumeSpecName "kube-api-access-khls2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.199279 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-xt5t7" podStartSLOduration=1.6380877090000001 podStartE2EDuration="10.199256802s" podCreationTimestamp="2026-02-20 08:23:44 +0000 UTC" firstStartedPulling="2026-02-20 08:23:45.30840186 +0000 UTC m=+1074.282896710" lastFinishedPulling="2026-02-20 08:23:53.869570983 +0000 UTC m=+1082.844065803" observedRunningTime="2026-02-20 08:23:54.186610499 +0000 UTC m=+1083.161105319" watchObservedRunningTime="2026-02-20 08:23:54.199256802 +0000 UTC m=+1083.173751622" Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.214405 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-config" (OuterVolumeSpecName: "config") pod "fd1b85e2-1b5f-4993-bc66-72daaf6aa79d" (UID: "fd1b85e2-1b5f-4993-bc66-72daaf6aa79d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.230842 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fd1b85e2-1b5f-4993-bc66-72daaf6aa79d" (UID: "fd1b85e2-1b5f-4993-bc66-72daaf6aa79d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.254407 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "fd1b85e2-1b5f-4993-bc66-72daaf6aa79d" (UID: "fd1b85e2-1b5f-4993-bc66-72daaf6aa79d"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.260773 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.260802 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khls2\" (UniqueName: \"kubernetes.io/projected/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-kube-api-access-khls2\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.260812 4948 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.260821 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.260830 4948 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.098983 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.125617 4948 generic.go:334] "Generic (PLEG): container finished" podID="8b1ee525-1a6e-461d-aa4d-178ab601ec9d" containerID="6a2830020d67b39ba1edbe811de2ebbd05b2b8a02d0578d95553d45d985544af" exitCode=0 Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.125676 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d46988dcb-rn6qs" event={"ID":"8b1ee525-1a6e-461d-aa4d-178ab601ec9d","Type":"ContainerDied","Data":"6a2830020d67b39ba1edbe811de2ebbd05b2b8a02d0578d95553d45d985544af"} Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.125704 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d46988dcb-rn6qs" event={"ID":"8b1ee525-1a6e-461d-aa4d-178ab601ec9d","Type":"ContainerDied","Data":"101e3a4bd62000e7fe0c14babf5f059dc5dad0bc38ccadaa7bd0301c2ad9ec7f"} Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.125720 4948 scope.go:117] "RemoveContainer" containerID="6a2830020d67b39ba1edbe811de2ebbd05b2b8a02d0578d95553d45d985544af" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.125836 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.146238 4948 generic.go:334] "Generic (PLEG): container finished" podID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerID="d2ec8efadba57da267b73b627d540acadced09eb54d33e0b4318c801f4b817ed" exitCode=2 Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.146267 4948 generic.go:334] "Generic (PLEG): container finished" podID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerID="8372790439c71a0ea7e4da810ca880dff1b44a6d3fa022b0a9cbbafbf8e763cc" exitCode=0 Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.146334 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0894350a-0df8-4c04-b7c7-fd8f93139959","Type":"ContainerStarted","Data":"c4e7730b1940fa1741db63557d39a1291cbd6edeee74b043d33b1623adece0d5"} Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.146359 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0894350a-0df8-4c04-b7c7-fd8f93139959","Type":"ContainerDied","Data":"d2ec8efadba57da267b73b627d540acadced09eb54d33e0b4318c801f4b817ed"} Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.146371 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0894350a-0df8-4c04-b7c7-fd8f93139959","Type":"ContainerDied","Data":"8372790439c71a0ea7e4da810ca880dff1b44a6d3fa022b0a9cbbafbf8e763cc"} Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.149305 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.149716 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b5c5cc45d-7dcbq" event={"ID":"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d","Type":"ContainerDied","Data":"4c4522a0ae94ffaed4b0ddc1a755c0498d0a11f95af5418408b70c0351708567"} Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.170210 4948 scope.go:117] "RemoveContainer" containerID="26cbea34a0a499d9db14acd1bf75d24dc0d0a677df528028730bacad671ebd26" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.176582 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-public-tls-certs\") pod \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.176650 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-internal-tls-certs\") pod \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.176695 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-logs\") pod \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.176775 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-scripts\") pod \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.176848 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bpmf\" (UniqueName: \"kubernetes.io/projected/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-kube-api-access-2bpmf\") pod \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.176937 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-combined-ca-bundle\") pod \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.176998 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-config-data\") pod \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.177876 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-logs" (OuterVolumeSpecName: "logs") pod "8b1ee525-1a6e-461d-aa4d-178ab601ec9d" (UID: "8b1ee525-1a6e-461d-aa4d-178ab601ec9d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.185490 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-scripts" (OuterVolumeSpecName: "scripts") pod "8b1ee525-1a6e-461d-aa4d-178ab601ec9d" (UID: "8b1ee525-1a6e-461d-aa4d-178ab601ec9d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.187892 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-b5c5cc45d-7dcbq"] Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.200081 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-kube-api-access-2bpmf" (OuterVolumeSpecName: "kube-api-access-2bpmf") pod "8b1ee525-1a6e-461d-aa4d-178ab601ec9d" (UID: "8b1ee525-1a6e-461d-aa4d-178ab601ec9d"). InnerVolumeSpecName "kube-api-access-2bpmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.206452 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-b5c5cc45d-7dcbq"] Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.269503 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8b1ee525-1a6e-461d-aa4d-178ab601ec9d" (UID: "8b1ee525-1a6e-461d-aa4d-178ab601ec9d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.282617 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.282642 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bpmf\" (UniqueName: \"kubernetes.io/projected/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-kube-api-access-2bpmf\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.282653 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.282662 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.333607 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-config-data" (OuterVolumeSpecName: "config-data") pod "8b1ee525-1a6e-461d-aa4d-178ab601ec9d" (UID: "8b1ee525-1a6e-461d-aa4d-178ab601ec9d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.338619 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8b1ee525-1a6e-461d-aa4d-178ab601ec9d" (UID: "8b1ee525-1a6e-461d-aa4d-178ab601ec9d"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.339113 4948 scope.go:117] "RemoveContainer" containerID="6a2830020d67b39ba1edbe811de2ebbd05b2b8a02d0578d95553d45d985544af" Feb 20 08:23:55 crc kubenswrapper[4948]: E0220 08:23:55.339870 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a2830020d67b39ba1edbe811de2ebbd05b2b8a02d0578d95553d45d985544af\": container with ID starting with 6a2830020d67b39ba1edbe811de2ebbd05b2b8a02d0578d95553d45d985544af not found: ID does not exist" containerID="6a2830020d67b39ba1edbe811de2ebbd05b2b8a02d0578d95553d45d985544af" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.339956 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a2830020d67b39ba1edbe811de2ebbd05b2b8a02d0578d95553d45d985544af"} err="failed to get container status \"6a2830020d67b39ba1edbe811de2ebbd05b2b8a02d0578d95553d45d985544af\": rpc error: code = NotFound desc = could not find container \"6a2830020d67b39ba1edbe811de2ebbd05b2b8a02d0578d95553d45d985544af\": container with ID starting with 6a2830020d67b39ba1edbe811de2ebbd05b2b8a02d0578d95553d45d985544af not found: ID does not exist" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.340078 4948 scope.go:117] "RemoveContainer" containerID="26cbea34a0a499d9db14acd1bf75d24dc0d0a677df528028730bacad671ebd26" Feb 20 08:23:55 crc kubenswrapper[4948]: E0220 08:23:55.344104 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26cbea34a0a499d9db14acd1bf75d24dc0d0a677df528028730bacad671ebd26\": container with ID starting with 26cbea34a0a499d9db14acd1bf75d24dc0d0a677df528028730bacad671ebd26 not found: ID does not exist" containerID="26cbea34a0a499d9db14acd1bf75d24dc0d0a677df528028730bacad671ebd26" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.344223 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26cbea34a0a499d9db14acd1bf75d24dc0d0a677df528028730bacad671ebd26"} err="failed to get container status \"26cbea34a0a499d9db14acd1bf75d24dc0d0a677df528028730bacad671ebd26\": rpc error: code = NotFound desc = could not find container \"26cbea34a0a499d9db14acd1bf75d24dc0d0a677df528028730bacad671ebd26\": container with ID starting with 26cbea34a0a499d9db14acd1bf75d24dc0d0a677df528028730bacad671ebd26 not found: ID does not exist" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.344294 4948 scope.go:117] "RemoveContainer" containerID="4fc7cff150a17fca1672ea7d9eae41c85a175b0d1f41224639307a4d7025cd12" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.346893 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "8b1ee525-1a6e-461d-aa4d-178ab601ec9d" (UID: "8b1ee525-1a6e-461d-aa4d-178ab601ec9d"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.371835 4948 scope.go:117] "RemoveContainer" containerID="fceb35ea8befbc40c2edb25ea5ef78a9c4a759e83b13e451735a1e2ccf6a99a8" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.384276 4948 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.384423 4948 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.384478 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.456906 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-d46988dcb-rn6qs"] Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.469518 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-d46988dcb-rn6qs"] Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.733074 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b1ee525-1a6e-461d-aa4d-178ab601ec9d" path="/var/lib/kubelet/pods/8b1ee525-1a6e-461d-aa4d-178ab601ec9d/volumes" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.733823 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd1b85e2-1b5f-4993-bc66-72daaf6aa79d" path="/var/lib/kubelet/pods/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d/volumes" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.157291 4948 generic.go:334] "Generic (PLEG): container finished" podID="d9399b9c-0935-4735-8b08-96d4a29d4ba8" containerID="656e84496ca041e8a454b07ed67e165b130af1030016564ae4519b8e2e6f11e5" exitCode=0 Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.157369 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d9399b9c-0935-4735-8b08-96d4a29d4ba8","Type":"ContainerDied","Data":"656e84496ca041e8a454b07ed67e165b130af1030016564ae4519b8e2e6f11e5"} Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.157555 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d9399b9c-0935-4735-8b08-96d4a29d4ba8","Type":"ContainerDied","Data":"fecc6864756af4b1a18a9eaf7ed69fde086ccbc2685b82cc1f598b9178cb31a0"} Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.157570 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fecc6864756af4b1a18a9eaf7ed69fde086ccbc2685b82cc1f598b9178cb31a0" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.165045 4948 generic.go:334] "Generic (PLEG): container finished" podID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerID="5a880cf140d0830077bc8b3d8a9a05a1bd68a08c91042ee55e74a3422f78c6e4" exitCode=0 Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.165079 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0894350a-0df8-4c04-b7c7-fd8f93139959","Type":"ContainerDied","Data":"5a880cf140d0830077bc8b3d8a9a05a1bd68a08c91042ee55e74a3422f78c6e4"} Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.181308 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.302243 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-public-tls-certs\") pod \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.302302 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdk7s\" (UniqueName: \"kubernetes.io/projected/d9399b9c-0935-4735-8b08-96d4a29d4ba8-kube-api-access-jdk7s\") pod \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.302345 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9399b9c-0935-4735-8b08-96d4a29d4ba8-logs\") pod \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.302405 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-combined-ca-bundle\") pod \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.302481 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d9399b9c-0935-4735-8b08-96d4a29d4ba8-httpd-run\") pod \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.302564 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.302625 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-config-data\") pod \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.302654 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-scripts\") pod \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.303364 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9399b9c-0935-4735-8b08-96d4a29d4ba8-logs" (OuterVolumeSpecName: "logs") pod "d9399b9c-0935-4735-8b08-96d4a29d4ba8" (UID: "d9399b9c-0935-4735-8b08-96d4a29d4ba8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.303386 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9399b9c-0935-4735-8b08-96d4a29d4ba8-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d9399b9c-0935-4735-8b08-96d4a29d4ba8" (UID: "d9399b9c-0935-4735-8b08-96d4a29d4ba8"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.306765 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9399b9c-0935-4735-8b08-96d4a29d4ba8-kube-api-access-jdk7s" (OuterVolumeSpecName: "kube-api-access-jdk7s") pod "d9399b9c-0935-4735-8b08-96d4a29d4ba8" (UID: "d9399b9c-0935-4735-8b08-96d4a29d4ba8"). InnerVolumeSpecName "kube-api-access-jdk7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.324115 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "d9399b9c-0935-4735-8b08-96d4a29d4ba8" (UID: "d9399b9c-0935-4735-8b08-96d4a29d4ba8"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.324200 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-scripts" (OuterVolumeSpecName: "scripts") pod "d9399b9c-0935-4735-8b08-96d4a29d4ba8" (UID: "d9399b9c-0935-4735-8b08-96d4a29d4ba8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.332938 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d9399b9c-0935-4735-8b08-96d4a29d4ba8" (UID: "d9399b9c-0935-4735-8b08-96d4a29d4ba8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.357626 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d9399b9c-0935-4735-8b08-96d4a29d4ba8" (UID: "d9399b9c-0935-4735-8b08-96d4a29d4ba8"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.393595 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-config-data" (OuterVolumeSpecName: "config-data") pod "d9399b9c-0935-4735-8b08-96d4a29d4ba8" (UID: "d9399b9c-0935-4735-8b08-96d4a29d4ba8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.405035 4948 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d9399b9c-0935-4735-8b08-96d4a29d4ba8-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.405105 4948 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.405118 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.405129 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.405140 4948 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.405232 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdk7s\" (UniqueName: \"kubernetes.io/projected/d9399b9c-0935-4735-8b08-96d4a29d4ba8-kube-api-access-jdk7s\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.405246 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9399b9c-0935-4735-8b08-96d4a29d4ba8-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.405257 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.425173 4948 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.506776 4948 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.903604 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.016940 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-internal-tls-certs\") pod \"33b35605-1057-4e64-99a7-a7273351d6c2\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.017017 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/33b35605-1057-4e64-99a7-a7273351d6c2-httpd-run\") pod \"33b35605-1057-4e64-99a7-a7273351d6c2\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.017045 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-combined-ca-bundle\") pod \"33b35605-1057-4e64-99a7-a7273351d6c2\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.017099 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-config-data\") pod \"33b35605-1057-4e64-99a7-a7273351d6c2\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.017114 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33b35605-1057-4e64-99a7-a7273351d6c2-logs\") pod \"33b35605-1057-4e64-99a7-a7273351d6c2\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.017134 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-scripts\") pod \"33b35605-1057-4e64-99a7-a7273351d6c2\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.017193 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6phw\" (UniqueName: \"kubernetes.io/projected/33b35605-1057-4e64-99a7-a7273351d6c2-kube-api-access-q6phw\") pod \"33b35605-1057-4e64-99a7-a7273351d6c2\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.017231 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"33b35605-1057-4e64-99a7-a7273351d6c2\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.019185 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33b35605-1057-4e64-99a7-a7273351d6c2-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "33b35605-1057-4e64-99a7-a7273351d6c2" (UID: "33b35605-1057-4e64-99a7-a7273351d6c2"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.027089 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33b35605-1057-4e64-99a7-a7273351d6c2-logs" (OuterVolumeSpecName: "logs") pod "33b35605-1057-4e64-99a7-a7273351d6c2" (UID: "33b35605-1057-4e64-99a7-a7273351d6c2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.037343 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "33b35605-1057-4e64-99a7-a7273351d6c2" (UID: "33b35605-1057-4e64-99a7-a7273351d6c2"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.039452 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-scripts" (OuterVolumeSpecName: "scripts") pod "33b35605-1057-4e64-99a7-a7273351d6c2" (UID: "33b35605-1057-4e64-99a7-a7273351d6c2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.066170 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33b35605-1057-4e64-99a7-a7273351d6c2-kube-api-access-q6phw" (OuterVolumeSpecName: "kube-api-access-q6phw") pod "33b35605-1057-4e64-99a7-a7273351d6c2" (UID: "33b35605-1057-4e64-99a7-a7273351d6c2"). InnerVolumeSpecName "kube-api-access-q6phw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.103795 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "33b35605-1057-4e64-99a7-a7273351d6c2" (UID: "33b35605-1057-4e64-99a7-a7273351d6c2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.118218 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-config-data" (OuterVolumeSpecName: "config-data") pod "33b35605-1057-4e64-99a7-a7273351d6c2" (UID: "33b35605-1057-4e64-99a7-a7273351d6c2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.118726 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-config-data\") pod \"33b35605-1057-4e64-99a7-a7273351d6c2\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " Feb 20 08:23:57 crc kubenswrapper[4948]: W0220 08:23:57.119183 4948 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/33b35605-1057-4e64-99a7-a7273351d6c2/volumes/kubernetes.io~secret/config-data Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.122708 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-config-data" (OuterVolumeSpecName: "config-data") pod "33b35605-1057-4e64-99a7-a7273351d6c2" (UID: "33b35605-1057-4e64-99a7-a7273351d6c2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.119913 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "33b35605-1057-4e64-99a7-a7273351d6c2" (UID: "33b35605-1057-4e64-99a7-a7273351d6c2"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.119333 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6phw\" (UniqueName: \"kubernetes.io/projected/33b35605-1057-4e64-99a7-a7273351d6c2-kube-api-access-q6phw\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.123144 4948 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.127625 4948 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/33b35605-1057-4e64-99a7-a7273351d6c2-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.127893 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.128021 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.128102 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33b35605-1057-4e64-99a7-a7273351d6c2-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.128158 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.145927 4948 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.180370 4948 generic.go:334] "Generic (PLEG): container finished" podID="33b35605-1057-4e64-99a7-a7273351d6c2" containerID="08ec8e2a4374c5a50916bcc6e4fbc0d7957aac63564273dd75ed243c1b2a0679" exitCode=0 Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.180502 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.182029 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.183020 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"33b35605-1057-4e64-99a7-a7273351d6c2","Type":"ContainerDied","Data":"08ec8e2a4374c5a50916bcc6e4fbc0d7957aac63564273dd75ed243c1b2a0679"} Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.183123 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"33b35605-1057-4e64-99a7-a7273351d6c2","Type":"ContainerDied","Data":"3abd35d213010021a69fde90dd757e9181aec2b84f4e7f71376ea0edef4300a5"} Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.183150 4948 scope.go:117] "RemoveContainer" containerID="08ec8e2a4374c5a50916bcc6e4fbc0d7957aac63564273dd75ed243c1b2a0679" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.221163 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.229882 4948 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.229940 4948 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.236334 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.246915 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.261105 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.284290 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:23:57 crc kubenswrapper[4948]: E0220 08:23:57.284646 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33b35605-1057-4e64-99a7-a7273351d6c2" containerName="glance-httpd" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.284658 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="33b35605-1057-4e64-99a7-a7273351d6c2" containerName="glance-httpd" Feb 20 08:23:57 crc kubenswrapper[4948]: E0220 08:23:57.284669 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd1b85e2-1b5f-4993-bc66-72daaf6aa79d" containerName="neutron-api" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.284676 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd1b85e2-1b5f-4993-bc66-72daaf6aa79d" containerName="neutron-api" Feb 20 08:23:57 crc kubenswrapper[4948]: E0220 08:23:57.284690 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b1ee525-1a6e-461d-aa4d-178ab601ec9d" containerName="placement-api" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.284697 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b1ee525-1a6e-461d-aa4d-178ab601ec9d" containerName="placement-api" Feb 20 08:23:57 crc kubenswrapper[4948]: E0220 08:23:57.284718 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd1b85e2-1b5f-4993-bc66-72daaf6aa79d" containerName="neutron-httpd" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.284724 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd1b85e2-1b5f-4993-bc66-72daaf6aa79d" containerName="neutron-httpd" Feb 20 08:23:57 crc kubenswrapper[4948]: E0220 08:23:57.284742 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b1ee525-1a6e-461d-aa4d-178ab601ec9d" containerName="placement-log" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.284747 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b1ee525-1a6e-461d-aa4d-178ab601ec9d" containerName="placement-log" Feb 20 08:23:57 crc kubenswrapper[4948]: E0220 08:23:57.284757 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9399b9c-0935-4735-8b08-96d4a29d4ba8" containerName="glance-log" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.284762 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9399b9c-0935-4735-8b08-96d4a29d4ba8" containerName="glance-log" Feb 20 08:23:57 crc kubenswrapper[4948]: E0220 08:23:57.284774 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33b35605-1057-4e64-99a7-a7273351d6c2" containerName="glance-log" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.284779 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="33b35605-1057-4e64-99a7-a7273351d6c2" containerName="glance-log" Feb 20 08:23:57 crc kubenswrapper[4948]: E0220 08:23:57.284792 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9399b9c-0935-4735-8b08-96d4a29d4ba8" containerName="glance-httpd" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.284799 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9399b9c-0935-4735-8b08-96d4a29d4ba8" containerName="glance-httpd" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.284989 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="33b35605-1057-4e64-99a7-a7273351d6c2" containerName="glance-httpd" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.285011 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b1ee525-1a6e-461d-aa4d-178ab601ec9d" containerName="placement-log" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.285020 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9399b9c-0935-4735-8b08-96d4a29d4ba8" containerName="glance-log" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.285031 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="33b35605-1057-4e64-99a7-a7273351d6c2" containerName="glance-log" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.285041 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd1b85e2-1b5f-4993-bc66-72daaf6aa79d" containerName="neutron-api" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.285047 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd1b85e2-1b5f-4993-bc66-72daaf6aa79d" containerName="neutron-httpd" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.285056 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b1ee525-1a6e-461d-aa4d-178ab601ec9d" containerName="placement-api" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.285067 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9399b9c-0935-4735-8b08-96d4a29d4ba8" containerName="glance-httpd" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.285858 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.286804 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.286879 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.287373 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.300658 4948 scope.go:117] "RemoveContainer" containerID="9232d9b3b9f362e1b01dbc32c5300081164bdbed5e40b2c4cd59a5e4f4b28166" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.308117 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.308351 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.308588 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.308716 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.308932 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.309242 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-bgxj8" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.331442 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-config-data\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.331474 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.331494 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fcb965e-4d04-4863-a966-39a83f458fa6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.331523 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4fcb965e-4d04-4863-a966-39a83f458fa6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.331541 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-scripts\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.331558 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.331578 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-logs\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.331607 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.331624 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4fcb965e-4d04-4863-a966-39a83f458fa6-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.331644 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.331688 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nx7wl\" (UniqueName: \"kubernetes.io/projected/4fcb965e-4d04-4863-a966-39a83f458fa6-kube-api-access-nx7wl\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.331708 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xl9fz\" (UniqueName: \"kubernetes.io/projected/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-kube-api-access-xl9fz\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.331731 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fcb965e-4d04-4863-a966-39a83f458fa6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.331758 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.331798 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4fcb965e-4d04-4863-a966-39a83f458fa6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.331818 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4fcb965e-4d04-4863-a966-39a83f458fa6-logs\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.357731 4948 scope.go:117] "RemoveContainer" containerID="08ec8e2a4374c5a50916bcc6e4fbc0d7957aac63564273dd75ed243c1b2a0679" Feb 20 08:23:57 crc kubenswrapper[4948]: E0220 08:23:57.358430 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08ec8e2a4374c5a50916bcc6e4fbc0d7957aac63564273dd75ed243c1b2a0679\": container with ID starting with 08ec8e2a4374c5a50916bcc6e4fbc0d7957aac63564273dd75ed243c1b2a0679 not found: ID does not exist" containerID="08ec8e2a4374c5a50916bcc6e4fbc0d7957aac63564273dd75ed243c1b2a0679" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.358492 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08ec8e2a4374c5a50916bcc6e4fbc0d7957aac63564273dd75ed243c1b2a0679"} err="failed to get container status \"08ec8e2a4374c5a50916bcc6e4fbc0d7957aac63564273dd75ed243c1b2a0679\": rpc error: code = NotFound desc = could not find container \"08ec8e2a4374c5a50916bcc6e4fbc0d7957aac63564273dd75ed243c1b2a0679\": container with ID starting with 08ec8e2a4374c5a50916bcc6e4fbc0d7957aac63564273dd75ed243c1b2a0679 not found: ID does not exist" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.358532 4948 scope.go:117] "RemoveContainer" containerID="9232d9b3b9f362e1b01dbc32c5300081164bdbed5e40b2c4cd59a5e4f4b28166" Feb 20 08:23:57 crc kubenswrapper[4948]: E0220 08:23:57.358861 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9232d9b3b9f362e1b01dbc32c5300081164bdbed5e40b2c4cd59a5e4f4b28166\": container with ID starting with 9232d9b3b9f362e1b01dbc32c5300081164bdbed5e40b2c4cd59a5e4f4b28166 not found: ID does not exist" containerID="9232d9b3b9f362e1b01dbc32c5300081164bdbed5e40b2c4cd59a5e4f4b28166" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.358900 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9232d9b3b9f362e1b01dbc32c5300081164bdbed5e40b2c4cd59a5e4f4b28166"} err="failed to get container status \"9232d9b3b9f362e1b01dbc32c5300081164bdbed5e40b2c4cd59a5e4f4b28166\": rpc error: code = NotFound desc = could not find container \"9232d9b3b9f362e1b01dbc32c5300081164bdbed5e40b2c4cd59a5e4f4b28166\": container with ID starting with 9232d9b3b9f362e1b01dbc32c5300081164bdbed5e40b2c4cd59a5e4f4b28166 not found: ID does not exist" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.382738 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.433778 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4fcb965e-4d04-4863-a966-39a83f458fa6-logs\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.433843 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-config-data\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.433870 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.433893 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fcb965e-4d04-4863-a966-39a83f458fa6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.433934 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4fcb965e-4d04-4863-a966-39a83f458fa6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.433956 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-scripts\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.434021 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.434046 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-logs\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.434073 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.434094 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4fcb965e-4d04-4863-a966-39a83f458fa6-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.434123 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.434193 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nx7wl\" (UniqueName: \"kubernetes.io/projected/4fcb965e-4d04-4863-a966-39a83f458fa6-kube-api-access-nx7wl\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.434218 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xl9fz\" (UniqueName: \"kubernetes.io/projected/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-kube-api-access-xl9fz\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.434261 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fcb965e-4d04-4863-a966-39a83f458fa6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.434304 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.434323 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.434340 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.434259 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4fcb965e-4d04-4863-a966-39a83f458fa6-logs\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.434675 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4fcb965e-4d04-4863-a966-39a83f458fa6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.434684 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-logs\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.435041 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4fcb965e-4d04-4863-a966-39a83f458fa6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.435442 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.438765 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fcb965e-4d04-4863-a966-39a83f458fa6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.439007 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.439998 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4fcb965e-4d04-4863-a966-39a83f458fa6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.440851 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-config-data\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.442108 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.442388 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4fcb965e-4d04-4863-a966-39a83f458fa6-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.453240 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fcb965e-4d04-4863-a966-39a83f458fa6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.453706 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-scripts\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.457529 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xl9fz\" (UniqueName: \"kubernetes.io/projected/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-kube-api-access-xl9fz\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.457922 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nx7wl\" (UniqueName: \"kubernetes.io/projected/4fcb965e-4d04-4863-a966-39a83f458fa6-kube-api-access-nx7wl\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.475173 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.486826 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.623237 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.637440 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.738709 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33b35605-1057-4e64-99a7-a7273351d6c2" path="/var/lib/kubelet/pods/33b35605-1057-4e64-99a7-a7273351d6c2/volumes" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.739904 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9399b9c-0935-4735-8b08-96d4a29d4ba8" path="/var/lib/kubelet/pods/d9399b9c-0935-4735-8b08-96d4a29d4ba8/volumes" Feb 20 08:23:58 crc kubenswrapper[4948]: I0220 08:23:58.206620 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 08:23:58 crc kubenswrapper[4948]: I0220 08:23:58.332536 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:23:58 crc kubenswrapper[4948]: W0220 08:23:58.342557 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podacd5bc1f_fee5_4ef1_95f9_5c4d11bbc54a.slice/crio-584190f9085e41c64b0b41271b19963599bd1689d31b9325f6e15ec1fa547bb8 WatchSource:0}: Error finding container 584190f9085e41c64b0b41271b19963599bd1689d31b9325f6e15ec1fa547bb8: Status 404 returned error can't find the container with id 584190f9085e41c64b0b41271b19963599bd1689d31b9325f6e15ec1fa547bb8 Feb 20 08:23:59 crc kubenswrapper[4948]: I0220 08:23:59.203165 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4fcb965e-4d04-4863-a966-39a83f458fa6","Type":"ContainerStarted","Data":"d76f6785864615fbf907238c357a5d56e0a2c4b94dbff43d8df27025c8acb1e5"} Feb 20 08:23:59 crc kubenswrapper[4948]: I0220 08:23:59.208787 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a","Type":"ContainerStarted","Data":"584190f9085e41c64b0b41271b19963599bd1689d31b9325f6e15ec1fa547bb8"} Feb 20 08:24:00 crc kubenswrapper[4948]: I0220 08:24:00.217290 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4fcb965e-4d04-4863-a966-39a83f458fa6","Type":"ContainerStarted","Data":"d4b2de7726d17bd80f2876aa816146886b795551e9fe8200d4594e8816986240"} Feb 20 08:24:00 crc kubenswrapper[4948]: I0220 08:24:00.217735 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4fcb965e-4d04-4863-a966-39a83f458fa6","Type":"ContainerStarted","Data":"faa90aa4e853b5ab666d9f1be43146101722dceedf35406ce1ae282e0648e2d0"} Feb 20 08:24:00 crc kubenswrapper[4948]: I0220 08:24:00.220134 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a","Type":"ContainerStarted","Data":"d5c3a0258f4fedb1c452016b5c3dfafb6d6a2a6693dd942cded6707663c562d1"} Feb 20 08:24:00 crc kubenswrapper[4948]: I0220 08:24:00.220157 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a","Type":"ContainerStarted","Data":"168c4e34c8fa1f4e36ab74b6fd468244e8fe2da745463fcd00309d5e31d08129"} Feb 20 08:24:00 crc kubenswrapper[4948]: I0220 08:24:00.239677 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.239660963 podStartE2EDuration="3.239660963s" podCreationTimestamp="2026-02-20 08:23:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:24:00.236395852 +0000 UTC m=+1089.210890662" watchObservedRunningTime="2026-02-20 08:24:00.239660963 +0000 UTC m=+1089.214155783" Feb 20 08:24:00 crc kubenswrapper[4948]: I0220 08:24:00.264589 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.2645723 podStartE2EDuration="3.2645723s" podCreationTimestamp="2026-02-20 08:23:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:24:00.259476044 +0000 UTC m=+1089.233970864" watchObservedRunningTime="2026-02-20 08:24:00.2645723 +0000 UTC m=+1089.239067120" Feb 20 08:24:06 crc kubenswrapper[4948]: I0220 08:24:06.272140 4948 generic.go:334] "Generic (PLEG): container finished" podID="3fdb2e02-8713-4205-ae31-0af33f8e4047" containerID="53a1cca1717eaf7448d7d84b4b68c8bad0f23b2d1cc07d696f16d15b417b8356" exitCode=0 Feb 20 08:24:06 crc kubenswrapper[4948]: I0220 08:24:06.272310 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xt5t7" event={"ID":"3fdb2e02-8713-4205-ae31-0af33f8e4047","Type":"ContainerDied","Data":"53a1cca1717eaf7448d7d84b4b68c8bad0f23b2d1cc07d696f16d15b417b8356"} Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.623930 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.624295 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.641265 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.641701 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.655346 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.665027 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.671596 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.692057 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.722896 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xt5t7" Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.845822 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-scripts\") pod \"3fdb2e02-8713-4205-ae31-0af33f8e4047\" (UID: \"3fdb2e02-8713-4205-ae31-0af33f8e4047\") " Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.845873 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-config-data\") pod \"3fdb2e02-8713-4205-ae31-0af33f8e4047\" (UID: \"3fdb2e02-8713-4205-ae31-0af33f8e4047\") " Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.846009 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j4jh4\" (UniqueName: \"kubernetes.io/projected/3fdb2e02-8713-4205-ae31-0af33f8e4047-kube-api-access-j4jh4\") pod \"3fdb2e02-8713-4205-ae31-0af33f8e4047\" (UID: \"3fdb2e02-8713-4205-ae31-0af33f8e4047\") " Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.846095 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-combined-ca-bundle\") pod \"3fdb2e02-8713-4205-ae31-0af33f8e4047\" (UID: \"3fdb2e02-8713-4205-ae31-0af33f8e4047\") " Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.853149 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-scripts" (OuterVolumeSpecName: "scripts") pod "3fdb2e02-8713-4205-ae31-0af33f8e4047" (UID: "3fdb2e02-8713-4205-ae31-0af33f8e4047"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.853171 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fdb2e02-8713-4205-ae31-0af33f8e4047-kube-api-access-j4jh4" (OuterVolumeSpecName: "kube-api-access-j4jh4") pod "3fdb2e02-8713-4205-ae31-0af33f8e4047" (UID: "3fdb2e02-8713-4205-ae31-0af33f8e4047"). InnerVolumeSpecName "kube-api-access-j4jh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.871606 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-config-data" (OuterVolumeSpecName: "config-data") pod "3fdb2e02-8713-4205-ae31-0af33f8e4047" (UID: "3fdb2e02-8713-4205-ae31-0af33f8e4047"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.872585 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3fdb2e02-8713-4205-ae31-0af33f8e4047" (UID: "3fdb2e02-8713-4205-ae31-0af33f8e4047"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.948025 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j4jh4\" (UniqueName: \"kubernetes.io/projected/3fdb2e02-8713-4205-ae31-0af33f8e4047-kube-api-access-j4jh4\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.948353 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.948363 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.948372 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.024691 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.024755 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.305543 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xt5t7" event={"ID":"3fdb2e02-8713-4205-ae31-0af33f8e4047","Type":"ContainerDied","Data":"d20f6a9cd2d1756d07c023cf4fc3e418531e5c20161aa674031ec3aa6a8df0d2"} Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.305587 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d20f6a9cd2d1756d07c023cf4fc3e418531e5c20161aa674031ec3aa6a8df0d2" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.305659 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xt5t7" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.306707 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.306932 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.307173 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.307931 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.389775 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 20 08:24:08 crc kubenswrapper[4948]: E0220 08:24:08.390301 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fdb2e02-8713-4205-ae31-0af33f8e4047" containerName="nova-cell0-conductor-db-sync" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.390325 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fdb2e02-8713-4205-ae31-0af33f8e4047" containerName="nova-cell0-conductor-db-sync" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.390587 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fdb2e02-8713-4205-ae31-0af33f8e4047" containerName="nova-cell0-conductor-db-sync" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.391467 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.396920 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-2tb2s" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.397379 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.400557 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.458432 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/873a6553-9637-4f2a-a743-f33cde7fc883-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"873a6553-9637-4f2a-a743-f33cde7fc883\") " pod="openstack/nova-cell0-conductor-0" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.458534 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/873a6553-9637-4f2a-a743-f33cde7fc883-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"873a6553-9637-4f2a-a743-f33cde7fc883\") " pod="openstack/nova-cell0-conductor-0" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.458641 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfd9z\" (UniqueName: \"kubernetes.io/projected/873a6553-9637-4f2a-a743-f33cde7fc883-kube-api-access-rfd9z\") pod \"nova-cell0-conductor-0\" (UID: \"873a6553-9637-4f2a-a743-f33cde7fc883\") " pod="openstack/nova-cell0-conductor-0" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.560382 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/873a6553-9637-4f2a-a743-f33cde7fc883-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"873a6553-9637-4f2a-a743-f33cde7fc883\") " pod="openstack/nova-cell0-conductor-0" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.560447 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfd9z\" (UniqueName: \"kubernetes.io/projected/873a6553-9637-4f2a-a743-f33cde7fc883-kube-api-access-rfd9z\") pod \"nova-cell0-conductor-0\" (UID: \"873a6553-9637-4f2a-a743-f33cde7fc883\") " pod="openstack/nova-cell0-conductor-0" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.560516 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/873a6553-9637-4f2a-a743-f33cde7fc883-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"873a6553-9637-4f2a-a743-f33cde7fc883\") " pod="openstack/nova-cell0-conductor-0" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.567020 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/873a6553-9637-4f2a-a743-f33cde7fc883-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"873a6553-9637-4f2a-a743-f33cde7fc883\") " pod="openstack/nova-cell0-conductor-0" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.571150 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/873a6553-9637-4f2a-a743-f33cde7fc883-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"873a6553-9637-4f2a-a743-f33cde7fc883\") " pod="openstack/nova-cell0-conductor-0" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.581415 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfd9z\" (UniqueName: \"kubernetes.io/projected/873a6553-9637-4f2a-a743-f33cde7fc883-kube-api-access-rfd9z\") pod \"nova-cell0-conductor-0\" (UID: \"873a6553-9637-4f2a-a743-f33cde7fc883\") " pod="openstack/nova-cell0-conductor-0" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.730389 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 20 08:24:09 crc kubenswrapper[4948]: I0220 08:24:09.203657 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 20 08:24:09 crc kubenswrapper[4948]: I0220 08:24:09.316725 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"873a6553-9637-4f2a-a743-f33cde7fc883","Type":"ContainerStarted","Data":"d84333201b58aeae0039550c4336cd843c3f79fd281d4d0e31148878aaed4cf4"} Feb 20 08:24:10 crc kubenswrapper[4948]: I0220 08:24:10.182317 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 20 08:24:10 crc kubenswrapper[4948]: I0220 08:24:10.231177 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 20 08:24:10 crc kubenswrapper[4948]: I0220 08:24:10.326918 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"873a6553-9637-4f2a-a743-f33cde7fc883","Type":"ContainerStarted","Data":"0ae670dd037318612ffa9a4d844e01a84b89b3a67099c7467316a31abc2e9825"} Feb 20 08:24:10 crc kubenswrapper[4948]: I0220 08:24:10.327320 4948 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 20 08:24:10 crc kubenswrapper[4948]: I0220 08:24:10.327333 4948 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 20 08:24:10 crc kubenswrapper[4948]: I0220 08:24:10.327905 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Feb 20 08:24:10 crc kubenswrapper[4948]: I0220 08:24:10.338571 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 20 08:24:10 crc kubenswrapper[4948]: I0220 08:24:10.344675 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.34465484 podStartE2EDuration="2.34465484s" podCreationTimestamp="2026-02-20 08:24:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:24:10.342669141 +0000 UTC m=+1099.317163971" watchObservedRunningTime="2026-02-20 08:24:10.34465484 +0000 UTC m=+1099.319149670" Feb 20 08:24:10 crc kubenswrapper[4948]: I0220 08:24:10.384456 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 20 08:24:13 crc kubenswrapper[4948]: I0220 08:24:13.339304 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Feb 20 08:24:18 crc kubenswrapper[4948]: I0220 08:24:18.757236 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.356206 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-qnlrn"] Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.358850 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qnlrn" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.363743 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.376338 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-qnlrn"] Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.385389 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.490195 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.491281 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.498759 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.504323 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-scripts\") pod \"nova-cell0-cell-mapping-qnlrn\" (UID: \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\") " pod="openstack/nova-cell0-cell-mapping-qnlrn" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.504419 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2j4xb\" (UniqueName: \"kubernetes.io/projected/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-kube-api-access-2j4xb\") pod \"nova-cell0-cell-mapping-qnlrn\" (UID: \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\") " pod="openstack/nova-cell0-cell-mapping-qnlrn" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.504507 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-config-data\") pod \"nova-cell0-cell-mapping-qnlrn\" (UID: \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\") " pod="openstack/nova-cell0-cell-mapping-qnlrn" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.504540 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-qnlrn\" (UID: \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\") " pod="openstack/nova-cell0-cell-mapping-qnlrn" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.507437 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.581584 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.591537 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.594241 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.598109 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.606129 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60006453-1770-44da-bd53-89701da2fa43-config-data\") pod \"nova-scheduler-0\" (UID: \"60006453-1770-44da-bd53-89701da2fa43\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.606193 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2j4xb\" (UniqueName: \"kubernetes.io/projected/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-kube-api-access-2j4xb\") pod \"nova-cell0-cell-mapping-qnlrn\" (UID: \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\") " pod="openstack/nova-cell0-cell-mapping-qnlrn" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.606265 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-config-data\") pod \"nova-cell0-cell-mapping-qnlrn\" (UID: \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\") " pod="openstack/nova-cell0-cell-mapping-qnlrn" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.606283 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60006453-1770-44da-bd53-89701da2fa43-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"60006453-1770-44da-bd53-89701da2fa43\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.606315 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-qnlrn\" (UID: \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\") " pod="openstack/nova-cell0-cell-mapping-qnlrn" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.606358 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-scripts\") pod \"nova-cell0-cell-mapping-qnlrn\" (UID: \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\") " pod="openstack/nova-cell0-cell-mapping-qnlrn" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.606388 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dl2cc\" (UniqueName: \"kubernetes.io/projected/60006453-1770-44da-bd53-89701da2fa43-kube-api-access-dl2cc\") pod \"nova-scheduler-0\" (UID: \"60006453-1770-44da-bd53-89701da2fa43\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.617898 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.620864 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-config-data\") pod \"nova-cell0-cell-mapping-qnlrn\" (UID: \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\") " pod="openstack/nova-cell0-cell-mapping-qnlrn" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.621546 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.626492 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-qnlrn\" (UID: \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\") " pod="openstack/nova-cell0-cell-mapping-qnlrn" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.640070 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.644583 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-scripts\") pod \"nova-cell0-cell-mapping-qnlrn\" (UID: \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\") " pod="openstack/nova-cell0-cell-mapping-qnlrn" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.646398 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2j4xb\" (UniqueName: \"kubernetes.io/projected/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-kube-api-access-2j4xb\") pod \"nova-cell0-cell-mapping-qnlrn\" (UID: \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\") " pod="openstack/nova-cell0-cell-mapping-qnlrn" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.701708 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qnlrn" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.709333 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad615fe7-b01a-4e4c-85b1-e028e5cdd866\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.709384 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dl2cc\" (UniqueName: \"kubernetes.io/projected/60006453-1770-44da-bd53-89701da2fa43-kube-api-access-dl2cc\") pod \"nova-scheduler-0\" (UID: \"60006453-1770-44da-bd53-89701da2fa43\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.709441 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blbc4\" (UniqueName: \"kubernetes.io/projected/025de664-c84b-4bb4-b953-13d95e9eee5d-kube-api-access-blbc4\") pod \"nova-api-0\" (UID: \"025de664-c84b-4bb4-b953-13d95e9eee5d\") " pod="openstack/nova-api-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.709461 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60006453-1770-44da-bd53-89701da2fa43-config-data\") pod \"nova-scheduler-0\" (UID: \"60006453-1770-44da-bd53-89701da2fa43\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.709501 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ll5q\" (UniqueName: \"kubernetes.io/projected/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-kube-api-access-2ll5q\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad615fe7-b01a-4e4c-85b1-e028e5cdd866\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.709521 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/025de664-c84b-4bb4-b953-13d95e9eee5d-config-data\") pod \"nova-api-0\" (UID: \"025de664-c84b-4bb4-b953-13d95e9eee5d\") " pod="openstack/nova-api-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.709569 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60006453-1770-44da-bd53-89701da2fa43-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"60006453-1770-44da-bd53-89701da2fa43\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.709600 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/025de664-c84b-4bb4-b953-13d95e9eee5d-logs\") pod \"nova-api-0\" (UID: \"025de664-c84b-4bb4-b953-13d95e9eee5d\") " pod="openstack/nova-api-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.709629 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/025de664-c84b-4bb4-b953-13d95e9eee5d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"025de664-c84b-4bb4-b953-13d95e9eee5d\") " pod="openstack/nova-api-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.709649 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad615fe7-b01a-4e4c-85b1-e028e5cdd866\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.715802 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.732553 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60006453-1770-44da-bd53-89701da2fa43-config-data\") pod \"nova-scheduler-0\" (UID: \"60006453-1770-44da-bd53-89701da2fa43\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.739680 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60006453-1770-44da-bd53-89701da2fa43-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"60006453-1770-44da-bd53-89701da2fa43\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.748681 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dl2cc\" (UniqueName: \"kubernetes.io/projected/60006453-1770-44da-bd53-89701da2fa43-kube-api-access-dl2cc\") pod \"nova-scheduler-0\" (UID: \"60006453-1770-44da-bd53-89701da2fa43\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.789212 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.795691 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.796009 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.799391 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.814289 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/025de664-c84b-4bb4-b953-13d95e9eee5d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"025de664-c84b-4bb4-b953-13d95e9eee5d\") " pod="openstack/nova-api-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.814346 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad615fe7-b01a-4e4c-85b1-e028e5cdd866\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.814395 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad615fe7-b01a-4e4c-85b1-e028e5cdd866\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.814468 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blbc4\" (UniqueName: \"kubernetes.io/projected/025de664-c84b-4bb4-b953-13d95e9eee5d-kube-api-access-blbc4\") pod \"nova-api-0\" (UID: \"025de664-c84b-4bb4-b953-13d95e9eee5d\") " pod="openstack/nova-api-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.814533 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ll5q\" (UniqueName: \"kubernetes.io/projected/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-kube-api-access-2ll5q\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad615fe7-b01a-4e4c-85b1-e028e5cdd866\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.814563 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/025de664-c84b-4bb4-b953-13d95e9eee5d-config-data\") pod \"nova-api-0\" (UID: \"025de664-c84b-4bb4-b953-13d95e9eee5d\") " pod="openstack/nova-api-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.814640 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/025de664-c84b-4bb4-b953-13d95e9eee5d-logs\") pod \"nova-api-0\" (UID: \"025de664-c84b-4bb4-b953-13d95e9eee5d\") " pod="openstack/nova-api-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.815198 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/025de664-c84b-4bb4-b953-13d95e9eee5d-logs\") pod \"nova-api-0\" (UID: \"025de664-c84b-4bb4-b953-13d95e9eee5d\") " pod="openstack/nova-api-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.824869 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad615fe7-b01a-4e4c-85b1-e028e5cdd866\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.833891 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/025de664-c84b-4bb4-b953-13d95e9eee5d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"025de664-c84b-4bb4-b953-13d95e9eee5d\") " pod="openstack/nova-api-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.834997 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad615fe7-b01a-4e4c-85b1-e028e5cdd866\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.836699 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/025de664-c84b-4bb4-b953-13d95e9eee5d-config-data\") pod \"nova-api-0\" (UID: \"025de664-c84b-4bb4-b953-13d95e9eee5d\") " pod="openstack/nova-api-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.841155 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.865889 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ll5q\" (UniqueName: \"kubernetes.io/projected/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-kube-api-access-2ll5q\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad615fe7-b01a-4e4c-85b1-e028e5cdd866\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.869653 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blbc4\" (UniqueName: \"kubernetes.io/projected/025de664-c84b-4bb4-b953-13d95e9eee5d-kube-api-access-blbc4\") pod \"nova-api-0\" (UID: \"025de664-c84b-4bb4-b953-13d95e9eee5d\") " pod="openstack/nova-api-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.911034 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-x22px"] Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.912604 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.915999 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-config-data\") pod \"nova-metadata-0\" (UID: \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\") " pod="openstack/nova-metadata-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.916092 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxxm8\" (UniqueName: \"kubernetes.io/projected/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-kube-api-access-vxxm8\") pod \"nova-metadata-0\" (UID: \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\") " pod="openstack/nova-metadata-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.916558 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\") " pod="openstack/nova-metadata-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.916611 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-logs\") pod \"nova-metadata-0\" (UID: \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\") " pod="openstack/nova-metadata-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.943519 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-x22px"] Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.948659 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.018873 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-config\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.018920 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-dns-svc\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.018960 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.019017 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxxm8\" (UniqueName: \"kubernetes.io/projected/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-kube-api-access-vxxm8\") pod \"nova-metadata-0\" (UID: \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\") " pod="openstack/nova-metadata-0" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.019074 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.019110 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hh28w\" (UniqueName: \"kubernetes.io/projected/538c1b6f-424d-48e3-a92b-7534fbf865fc-kube-api-access-hh28w\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.019133 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.019175 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\") " pod="openstack/nova-metadata-0" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.019208 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-logs\") pod \"nova-metadata-0\" (UID: \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\") " pod="openstack/nova-metadata-0" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.019245 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-config-data\") pod \"nova-metadata-0\" (UID: \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\") " pod="openstack/nova-metadata-0" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.020002 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-logs\") pod \"nova-metadata-0\" (UID: \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\") " pod="openstack/nova-metadata-0" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.024732 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\") " pod="openstack/nova-metadata-0" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.025299 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-config-data\") pod \"nova-metadata-0\" (UID: \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\") " pod="openstack/nova-metadata-0" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.040000 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxxm8\" (UniqueName: \"kubernetes.io/projected/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-kube-api-access-vxxm8\") pod \"nova-metadata-0\" (UID: \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\") " pod="openstack/nova-metadata-0" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.121404 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.121758 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hh28w\" (UniqueName: \"kubernetes.io/projected/538c1b6f-424d-48e3-a92b-7534fbf865fc-kube-api-access-hh28w\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.121950 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.122178 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-config\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.122258 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.122519 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-dns-svc\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.122556 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.122913 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-config\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.122919 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.123460 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-dns-svc\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.125418 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.139454 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hh28w\" (UniqueName: \"kubernetes.io/projected/538c1b6f-424d-48e3-a92b-7534fbf865fc-kube-api-access-hh28w\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.158644 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.205185 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.259383 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.412072 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-qnlrn"] Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.475878 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qnlrn" event={"ID":"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b","Type":"ContainerStarted","Data":"12945171d0e3aefdd46fee90ddbf80a6cc5d07d1053f4e8bc8f96de548e35fcc"} Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.493111 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.586823 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.616000 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-jl8ck"] Feb 20 08:24:20 crc kubenswrapper[4948]: W0220 08:24:20.616669 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1eea9d32_4e2a_4ad9_a402_7d7a483a3dff.slice/crio-4ce5c5ab612acce40e4a349276cab71f6a92423acc9a5a2deff1b1f12271b01e WatchSource:0}: Error finding container 4ce5c5ab612acce40e4a349276cab71f6a92423acc9a5a2deff1b1f12271b01e: Status 404 returned error can't find the container with id 4ce5c5ab612acce40e4a349276cab71f6a92423acc9a5a2deff1b1f12271b01e Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.617667 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-jl8ck" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.619883 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.620053 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.635685 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.646163 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-jl8ck"] Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.674752 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.747358 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-config-data\") pod \"nova-cell1-conductor-db-sync-jl8ck\" (UID: \"93f8e582-bd6a-44d4-a203-ba9950efcada\") " pod="openstack/nova-cell1-conductor-db-sync-jl8ck" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.747433 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-scripts\") pod \"nova-cell1-conductor-db-sync-jl8ck\" (UID: \"93f8e582-bd6a-44d4-a203-ba9950efcada\") " pod="openstack/nova-cell1-conductor-db-sync-jl8ck" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.747553 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-jl8ck\" (UID: \"93f8e582-bd6a-44d4-a203-ba9950efcada\") " pod="openstack/nova-cell1-conductor-db-sync-jl8ck" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.747619 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czd6x\" (UniqueName: \"kubernetes.io/projected/93f8e582-bd6a-44d4-a203-ba9950efcada-kube-api-access-czd6x\") pod \"nova-cell1-conductor-db-sync-jl8ck\" (UID: \"93f8e582-bd6a-44d4-a203-ba9950efcada\") " pod="openstack/nova-cell1-conductor-db-sync-jl8ck" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.850241 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-config-data\") pod \"nova-cell1-conductor-db-sync-jl8ck\" (UID: \"93f8e582-bd6a-44d4-a203-ba9950efcada\") " pod="openstack/nova-cell1-conductor-db-sync-jl8ck" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.850305 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-scripts\") pod \"nova-cell1-conductor-db-sync-jl8ck\" (UID: \"93f8e582-bd6a-44d4-a203-ba9950efcada\") " pod="openstack/nova-cell1-conductor-db-sync-jl8ck" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.850326 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-jl8ck\" (UID: \"93f8e582-bd6a-44d4-a203-ba9950efcada\") " pod="openstack/nova-cell1-conductor-db-sync-jl8ck" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.850350 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czd6x\" (UniqueName: \"kubernetes.io/projected/93f8e582-bd6a-44d4-a203-ba9950efcada-kube-api-access-czd6x\") pod \"nova-cell1-conductor-db-sync-jl8ck\" (UID: \"93f8e582-bd6a-44d4-a203-ba9950efcada\") " pod="openstack/nova-cell1-conductor-db-sync-jl8ck" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.856562 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-config-data\") pod \"nova-cell1-conductor-db-sync-jl8ck\" (UID: \"93f8e582-bd6a-44d4-a203-ba9950efcada\") " pod="openstack/nova-cell1-conductor-db-sync-jl8ck" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.857316 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-scripts\") pod \"nova-cell1-conductor-db-sync-jl8ck\" (UID: \"93f8e582-bd6a-44d4-a203-ba9950efcada\") " pod="openstack/nova-cell1-conductor-db-sync-jl8ck" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.863460 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-jl8ck\" (UID: \"93f8e582-bd6a-44d4-a203-ba9950efcada\") " pod="openstack/nova-cell1-conductor-db-sync-jl8ck" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.865285 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czd6x\" (UniqueName: \"kubernetes.io/projected/93f8e582-bd6a-44d4-a203-ba9950efcada-kube-api-access-czd6x\") pod \"nova-cell1-conductor-db-sync-jl8ck\" (UID: \"93f8e582-bd6a-44d4-a203-ba9950efcada\") " pod="openstack/nova-cell1-conductor-db-sync-jl8ck" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.929267 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-x22px"] Feb 20 08:24:21 crc kubenswrapper[4948]: I0220 08:24:21.053715 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-jl8ck" Feb 20 08:24:21 crc kubenswrapper[4948]: I0220 08:24:21.488062 4948 generic.go:334] "Generic (PLEG): container finished" podID="538c1b6f-424d-48e3-a92b-7534fbf865fc" containerID="70074ef7c3184ccb90626f2f932107dda3ac2287a8b654203232efa428fd6294" exitCode=0 Feb 20 08:24:21 crc kubenswrapper[4948]: I0220 08:24:21.488169 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-x22px" event={"ID":"538c1b6f-424d-48e3-a92b-7534fbf865fc","Type":"ContainerDied","Data":"70074ef7c3184ccb90626f2f932107dda3ac2287a8b654203232efa428fd6294"} Feb 20 08:24:21 crc kubenswrapper[4948]: I0220 08:24:21.488405 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-x22px" event={"ID":"538c1b6f-424d-48e3-a92b-7534fbf865fc","Type":"ContainerStarted","Data":"ca84f60c8f59ce30814af7db2ac25fea0b071bd0c109ef4bdad869f451268b2b"} Feb 20 08:24:21 crc kubenswrapper[4948]: I0220 08:24:21.491794 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ad615fe7-b01a-4e4c-85b1-e028e5cdd866","Type":"ContainerStarted","Data":"2e2fb5966e87b5f5075085440ad9aaba0868d3b137c6af57f293094a97faf543"} Feb 20 08:24:21 crc kubenswrapper[4948]: I0220 08:24:21.500252 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff","Type":"ContainerStarted","Data":"4ce5c5ab612acce40e4a349276cab71f6a92423acc9a5a2deff1b1f12271b01e"} Feb 20 08:24:21 crc kubenswrapper[4948]: I0220 08:24:21.503182 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"025de664-c84b-4bb4-b953-13d95e9eee5d","Type":"ContainerStarted","Data":"a2ceacf4091ad26a00570ff4d1bfb8fcf3b35d5c7789387f917c60bdba3eb6b8"} Feb 20 08:24:21 crc kubenswrapper[4948]: I0220 08:24:21.515998 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"60006453-1770-44da-bd53-89701da2fa43","Type":"ContainerStarted","Data":"22ac2dae6a489cbbc45577f9a64760c4218b1a60e308183ad1c3246e8923567d"} Feb 20 08:24:21 crc kubenswrapper[4948]: I0220 08:24:21.520552 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qnlrn" event={"ID":"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b","Type":"ContainerStarted","Data":"9a9d495cbe6ddcf0246661f3edacc2b8afe21e8ef43eca82e9140a67f21f1c6d"} Feb 20 08:24:21 crc kubenswrapper[4948]: I0220 08:24:21.536389 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-qnlrn" podStartSLOduration=2.53637047 podStartE2EDuration="2.53637047s" podCreationTimestamp="2026-02-20 08:24:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:24:21.53596579 +0000 UTC m=+1110.510460600" watchObservedRunningTime="2026-02-20 08:24:21.53637047 +0000 UTC m=+1110.510865290" Feb 20 08:24:21 crc kubenswrapper[4948]: I0220 08:24:21.557808 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-jl8ck"] Feb 20 08:24:21 crc kubenswrapper[4948]: W0220 08:24:21.562602 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93f8e582_bd6a_44d4_a203_ba9950efcada.slice/crio-bc36e2f68570b6936d5e5dbd27c44aba36cbc45ceca2d3d6a770bf8b08d1d326 WatchSource:0}: Error finding container bc36e2f68570b6936d5e5dbd27c44aba36cbc45ceca2d3d6a770bf8b08d1d326: Status 404 returned error can't find the container with id bc36e2f68570b6936d5e5dbd27c44aba36cbc45ceca2d3d6a770bf8b08d1d326 Feb 20 08:24:22 crc kubenswrapper[4948]: I0220 08:24:22.529632 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-jl8ck" event={"ID":"93f8e582-bd6a-44d4-a203-ba9950efcada","Type":"ContainerStarted","Data":"67cb9d3659cf07e0efb010d75eb62a0dddad421b940140c438f0a0714676a1ed"} Feb 20 08:24:22 crc kubenswrapper[4948]: I0220 08:24:22.529920 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-jl8ck" event={"ID":"93f8e582-bd6a-44d4-a203-ba9950efcada","Type":"ContainerStarted","Data":"bc36e2f68570b6936d5e5dbd27c44aba36cbc45ceca2d3d6a770bf8b08d1d326"} Feb 20 08:24:22 crc kubenswrapper[4948]: I0220 08:24:22.533651 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-x22px" event={"ID":"538c1b6f-424d-48e3-a92b-7534fbf865fc","Type":"ContainerStarted","Data":"3a9357bad2b3dca80dc83315104b35d29a1b980c7594ca0611bf4b94b36eb262"} Feb 20 08:24:22 crc kubenswrapper[4948]: I0220 08:24:22.548237 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-jl8ck" podStartSLOduration=2.548220019 podStartE2EDuration="2.548220019s" podCreationTimestamp="2026-02-20 08:24:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:24:22.545655567 +0000 UTC m=+1111.520150397" watchObservedRunningTime="2026-02-20 08:24:22.548220019 +0000 UTC m=+1111.522714829" Feb 20 08:24:22 crc kubenswrapper[4948]: I0220 08:24:22.595543 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-757b4f8459-x22px" podStartSLOduration=3.5955207 podStartE2EDuration="3.5955207s" podCreationTimestamp="2026-02-20 08:24:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:24:22.578844838 +0000 UTC m=+1111.553339658" watchObservedRunningTime="2026-02-20 08:24:22.5955207 +0000 UTC m=+1111.570015520" Feb 20 08:24:23 crc kubenswrapper[4948]: I0220 08:24:23.481022 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:24:23 crc kubenswrapper[4948]: I0220 08:24:23.493648 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 20 08:24:23 crc kubenswrapper[4948]: I0220 08:24:23.541271 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.479882 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.534308 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0894350a-0df8-4c04-b7c7-fd8f93139959-run-httpd\") pod \"0894350a-0df8-4c04-b7c7-fd8f93139959\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.534348 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-combined-ca-bundle\") pod \"0894350a-0df8-4c04-b7c7-fd8f93139959\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.534394 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txt8k\" (UniqueName: \"kubernetes.io/projected/0894350a-0df8-4c04-b7c7-fd8f93139959-kube-api-access-txt8k\") pod \"0894350a-0df8-4c04-b7c7-fd8f93139959\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.534416 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-scripts\") pod \"0894350a-0df8-4c04-b7c7-fd8f93139959\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.534453 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-config-data\") pod \"0894350a-0df8-4c04-b7c7-fd8f93139959\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.534473 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-sg-core-conf-yaml\") pod \"0894350a-0df8-4c04-b7c7-fd8f93139959\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.534581 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0894350a-0df8-4c04-b7c7-fd8f93139959-log-httpd\") pod \"0894350a-0df8-4c04-b7c7-fd8f93139959\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.535252 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0894350a-0df8-4c04-b7c7-fd8f93139959-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0894350a-0df8-4c04-b7c7-fd8f93139959" (UID: "0894350a-0df8-4c04-b7c7-fd8f93139959"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.535489 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0894350a-0df8-4c04-b7c7-fd8f93139959-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0894350a-0df8-4c04-b7c7-fd8f93139959" (UID: "0894350a-0df8-4c04-b7c7-fd8f93139959"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.540242 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0894350a-0df8-4c04-b7c7-fd8f93139959-kube-api-access-txt8k" (OuterVolumeSpecName: "kube-api-access-txt8k") pod "0894350a-0df8-4c04-b7c7-fd8f93139959" (UID: "0894350a-0df8-4c04-b7c7-fd8f93139959"). InnerVolumeSpecName "kube-api-access-txt8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.541544 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-scripts" (OuterVolumeSpecName: "scripts") pod "0894350a-0df8-4c04-b7c7-fd8f93139959" (UID: "0894350a-0df8-4c04-b7c7-fd8f93139959"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.552135 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"025de664-c84b-4bb4-b953-13d95e9eee5d","Type":"ContainerStarted","Data":"a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7"} Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.552176 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"025de664-c84b-4bb4-b953-13d95e9eee5d","Type":"ContainerStarted","Data":"2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56"} Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.553528 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"60006453-1770-44da-bd53-89701da2fa43","Type":"ContainerStarted","Data":"789d33d79de6243b08689af1b9cf996bdaa6473c8e9a5246cde2871757156c6a"} Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.558477 4948 generic.go:334] "Generic (PLEG): container finished" podID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerID="c4e7730b1940fa1741db63557d39a1291cbd6edeee74b043d33b1623adece0d5" exitCode=137 Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.558540 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0894350a-0df8-4c04-b7c7-fd8f93139959","Type":"ContainerDied","Data":"c4e7730b1940fa1741db63557d39a1291cbd6edeee74b043d33b1623adece0d5"} Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.558561 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0894350a-0df8-4c04-b7c7-fd8f93139959","Type":"ContainerDied","Data":"6657eaf082a861ad8f0f29c4c00aaf35d258d94004b82734a54c5722b14133f5"} Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.558581 4948 scope.go:117] "RemoveContainer" containerID="c4e7730b1940fa1741db63557d39a1291cbd6edeee74b043d33b1623adece0d5" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.558700 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.565607 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ad615fe7-b01a-4e4c-85b1-e028e5cdd866","Type":"ContainerStarted","Data":"2242b79213417225ca0b9ded36fc0195241950291c46926c9dff4400bf7ca4e9"} Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.565721 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="ad615fe7-b01a-4e4c-85b1-e028e5cdd866" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://2242b79213417225ca0b9ded36fc0195241950291c46926c9dff4400bf7ca4e9" gracePeriod=30 Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.576401 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0894350a-0df8-4c04-b7c7-fd8f93139959" (UID: "0894350a-0df8-4c04-b7c7-fd8f93139959"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.576576 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1eea9d32-4e2a-4ad9-a402-7d7a483a3dff" containerName="nova-metadata-log" containerID="cri-o://b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a" gracePeriod=30 Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.576734 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff","Type":"ContainerStarted","Data":"3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911"} Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.576773 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff","Type":"ContainerStarted","Data":"b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a"} Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.576848 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1eea9d32-4e2a-4ad9-a402-7d7a483a3dff" containerName="nova-metadata-metadata" containerID="cri-o://3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911" gracePeriod=30 Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.581777 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.580605578 podStartE2EDuration="5.581764532s" podCreationTimestamp="2026-02-20 08:24:19 +0000 UTC" firstStartedPulling="2026-02-20 08:24:20.590395671 +0000 UTC m=+1109.564907552" lastFinishedPulling="2026-02-20 08:24:23.591571686 +0000 UTC m=+1112.566066506" observedRunningTime="2026-02-20 08:24:24.572081108 +0000 UTC m=+1113.546575928" watchObservedRunningTime="2026-02-20 08:24:24.581764532 +0000 UTC m=+1113.556259342" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.593282 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.69013333 podStartE2EDuration="5.593262209s" podCreationTimestamp="2026-02-20 08:24:19 +0000 UTC" firstStartedPulling="2026-02-20 08:24:20.688467548 +0000 UTC m=+1109.662962368" lastFinishedPulling="2026-02-20 08:24:23.591596427 +0000 UTC m=+1112.566091247" observedRunningTime="2026-02-20 08:24:24.583709209 +0000 UTC m=+1113.558204019" watchObservedRunningTime="2026-02-20 08:24:24.593262209 +0000 UTC m=+1113.567757029" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.595804 4948 scope.go:117] "RemoveContainer" containerID="d2ec8efadba57da267b73b627d540acadced09eb54d33e0b4318c801f4b817ed" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.624736 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.55623708 podStartE2EDuration="5.624716299s" podCreationTimestamp="2026-02-20 08:24:19 +0000 UTC" firstStartedPulling="2026-02-20 08:24:20.525205938 +0000 UTC m=+1109.499700758" lastFinishedPulling="2026-02-20 08:24:23.593685157 +0000 UTC m=+1112.568179977" observedRunningTime="2026-02-20 08:24:24.622475454 +0000 UTC m=+1113.596970274" watchObservedRunningTime="2026-02-20 08:24:24.624716299 +0000 UTC m=+1113.599211119" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.639608 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txt8k\" (UniqueName: \"kubernetes.io/projected/0894350a-0df8-4c04-b7c7-fd8f93139959-kube-api-access-txt8k\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.639948 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.639959 4948 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.640008 4948 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0894350a-0df8-4c04-b7c7-fd8f93139959-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.640021 4948 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0894350a-0df8-4c04-b7c7-fd8f93139959-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.643250 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.6689121289999997 podStartE2EDuration="5.643240796s" podCreationTimestamp="2026-02-20 08:24:19 +0000 UTC" firstStartedPulling="2026-02-20 08:24:20.62182018 +0000 UTC m=+1109.596314990" lastFinishedPulling="2026-02-20 08:24:23.596148837 +0000 UTC m=+1112.570643657" observedRunningTime="2026-02-20 08:24:24.642225641 +0000 UTC m=+1113.616720461" watchObservedRunningTime="2026-02-20 08:24:24.643240796 +0000 UTC m=+1113.617735616" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.656924 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0894350a-0df8-4c04-b7c7-fd8f93139959" (UID: "0894350a-0df8-4c04-b7c7-fd8f93139959"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.691227 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-config-data" (OuterVolumeSpecName: "config-data") pod "0894350a-0df8-4c04-b7c7-fd8f93139959" (UID: "0894350a-0df8-4c04-b7c7-fd8f93139959"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.741250 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.741285 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.845454 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.886826 4948 scope.go:117] "RemoveContainer" containerID="8372790439c71a0ea7e4da810ca880dff1b44a6d3fa022b0a9cbbafbf8e763cc" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.906848 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.922322 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.940363 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:24:24 crc kubenswrapper[4948]: E0220 08:24:24.940706 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerName="ceilometer-notification-agent" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.940718 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerName="ceilometer-notification-agent" Feb 20 08:24:24 crc kubenswrapper[4948]: E0220 08:24:24.940743 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerName="ceilometer-central-agent" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.940750 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerName="ceilometer-central-agent" Feb 20 08:24:24 crc kubenswrapper[4948]: E0220 08:24:24.940760 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerName="sg-core" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.940766 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerName="sg-core" Feb 20 08:24:24 crc kubenswrapper[4948]: E0220 08:24:24.940779 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerName="proxy-httpd" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.940784 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerName="proxy-httpd" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.943774 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerName="ceilometer-central-agent" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.943800 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerName="ceilometer-notification-agent" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.943813 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerName="sg-core" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.943835 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerName="proxy-httpd" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.958321 4948 scope.go:117] "RemoveContainer" containerID="5a880cf140d0830077bc8b3d8a9a05a1bd68a08c91042ee55e74a3422f78c6e4" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.959540 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.966946 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.967724 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.986488 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.019397 4948 scope.go:117] "RemoveContainer" containerID="c4e7730b1940fa1741db63557d39a1291cbd6edeee74b043d33b1623adece0d5" Feb 20 08:24:25 crc kubenswrapper[4948]: E0220 08:24:25.022613 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4e7730b1940fa1741db63557d39a1291cbd6edeee74b043d33b1623adece0d5\": container with ID starting with c4e7730b1940fa1741db63557d39a1291cbd6edeee74b043d33b1623adece0d5 not found: ID does not exist" containerID="c4e7730b1940fa1741db63557d39a1291cbd6edeee74b043d33b1623adece0d5" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.022658 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4e7730b1940fa1741db63557d39a1291cbd6edeee74b043d33b1623adece0d5"} err="failed to get container status \"c4e7730b1940fa1741db63557d39a1291cbd6edeee74b043d33b1623adece0d5\": rpc error: code = NotFound desc = could not find container \"c4e7730b1940fa1741db63557d39a1291cbd6edeee74b043d33b1623adece0d5\": container with ID starting with c4e7730b1940fa1741db63557d39a1291cbd6edeee74b043d33b1623adece0d5 not found: ID does not exist" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.022682 4948 scope.go:117] "RemoveContainer" containerID="d2ec8efadba57da267b73b627d540acadced09eb54d33e0b4318c801f4b817ed" Feb 20 08:24:25 crc kubenswrapper[4948]: E0220 08:24:25.026100 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2ec8efadba57da267b73b627d540acadced09eb54d33e0b4318c801f4b817ed\": container with ID starting with d2ec8efadba57da267b73b627d540acadced09eb54d33e0b4318c801f4b817ed not found: ID does not exist" containerID="d2ec8efadba57da267b73b627d540acadced09eb54d33e0b4318c801f4b817ed" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.026141 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2ec8efadba57da267b73b627d540acadced09eb54d33e0b4318c801f4b817ed"} err="failed to get container status \"d2ec8efadba57da267b73b627d540acadced09eb54d33e0b4318c801f4b817ed\": rpc error: code = NotFound desc = could not find container \"d2ec8efadba57da267b73b627d540acadced09eb54d33e0b4318c801f4b817ed\": container with ID starting with d2ec8efadba57da267b73b627d540acadced09eb54d33e0b4318c801f4b817ed not found: ID does not exist" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.026168 4948 scope.go:117] "RemoveContainer" containerID="8372790439c71a0ea7e4da810ca880dff1b44a6d3fa022b0a9cbbafbf8e763cc" Feb 20 08:24:25 crc kubenswrapper[4948]: E0220 08:24:25.027659 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8372790439c71a0ea7e4da810ca880dff1b44a6d3fa022b0a9cbbafbf8e763cc\": container with ID starting with 8372790439c71a0ea7e4da810ca880dff1b44a6d3fa022b0a9cbbafbf8e763cc not found: ID does not exist" containerID="8372790439c71a0ea7e4da810ca880dff1b44a6d3fa022b0a9cbbafbf8e763cc" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.027683 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8372790439c71a0ea7e4da810ca880dff1b44a6d3fa022b0a9cbbafbf8e763cc"} err="failed to get container status \"8372790439c71a0ea7e4da810ca880dff1b44a6d3fa022b0a9cbbafbf8e763cc\": rpc error: code = NotFound desc = could not find container \"8372790439c71a0ea7e4da810ca880dff1b44a6d3fa022b0a9cbbafbf8e763cc\": container with ID starting with 8372790439c71a0ea7e4da810ca880dff1b44a6d3fa022b0a9cbbafbf8e763cc not found: ID does not exist" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.027712 4948 scope.go:117] "RemoveContainer" containerID="5a880cf140d0830077bc8b3d8a9a05a1bd68a08c91042ee55e74a3422f78c6e4" Feb 20 08:24:25 crc kubenswrapper[4948]: E0220 08:24:25.044169 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a880cf140d0830077bc8b3d8a9a05a1bd68a08c91042ee55e74a3422f78c6e4\": container with ID starting with 5a880cf140d0830077bc8b3d8a9a05a1bd68a08c91042ee55e74a3422f78c6e4 not found: ID does not exist" containerID="5a880cf140d0830077bc8b3d8a9a05a1bd68a08c91042ee55e74a3422f78c6e4" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.044215 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a880cf140d0830077bc8b3d8a9a05a1bd68a08c91042ee55e74a3422f78c6e4"} err="failed to get container status \"5a880cf140d0830077bc8b3d8a9a05a1bd68a08c91042ee55e74a3422f78c6e4\": rpc error: code = NotFound desc = could not find container \"5a880cf140d0830077bc8b3d8a9a05a1bd68a08c91042ee55e74a3422f78c6e4\": container with ID starting with 5a880cf140d0830077bc8b3d8a9a05a1bd68a08c91042ee55e74a3422f78c6e4 not found: ID does not exist" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.063592 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-scripts\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.063647 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee88c030-4657-480c-99a0-90ee4c2b247c-run-httpd\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.063672 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-config-data\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.063704 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrzbr\" (UniqueName: \"kubernetes.io/projected/ee88c030-4657-480c-99a0-90ee4c2b247c-kube-api-access-rrzbr\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.063732 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee88c030-4657-480c-99a0-90ee4c2b247c-log-httpd\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.063764 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.063816 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.111469 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.159534 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.165426 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-combined-ca-bundle\") pod \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\" (UID: \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\") " Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.165523 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-config-data\") pod \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\" (UID: \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\") " Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.165603 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxxm8\" (UniqueName: \"kubernetes.io/projected/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-kube-api-access-vxxm8\") pod \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\" (UID: \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\") " Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.165689 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-logs\") pod \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\" (UID: \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\") " Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.165882 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.165966 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-scripts\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.166000 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee88c030-4657-480c-99a0-90ee4c2b247c-run-httpd\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.166021 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-config-data\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.166046 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrzbr\" (UniqueName: \"kubernetes.io/projected/ee88c030-4657-480c-99a0-90ee4c2b247c-kube-api-access-rrzbr\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.166072 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee88c030-4657-480c-99a0-90ee4c2b247c-log-httpd\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.166099 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.167545 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee88c030-4657-480c-99a0-90ee4c2b247c-run-httpd\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.167571 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee88c030-4657-480c-99a0-90ee4c2b247c-log-httpd\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.167772 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-logs" (OuterVolumeSpecName: "logs") pod "1eea9d32-4e2a-4ad9-a402-7d7a483a3dff" (UID: "1eea9d32-4e2a-4ad9-a402-7d7a483a3dff"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.171230 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.171620 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-config-data\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.172166 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.172938 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-scripts\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.174190 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-kube-api-access-vxxm8" (OuterVolumeSpecName: "kube-api-access-vxxm8") pod "1eea9d32-4e2a-4ad9-a402-7d7a483a3dff" (UID: "1eea9d32-4e2a-4ad9-a402-7d7a483a3dff"). InnerVolumeSpecName "kube-api-access-vxxm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.191611 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrzbr\" (UniqueName: \"kubernetes.io/projected/ee88c030-4657-480c-99a0-90ee4c2b247c-kube-api-access-rrzbr\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.195790 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1eea9d32-4e2a-4ad9-a402-7d7a483a3dff" (UID: "1eea9d32-4e2a-4ad9-a402-7d7a483a3dff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.201272 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-config-data" (OuterVolumeSpecName: "config-data") pod "1eea9d32-4e2a-4ad9-a402-7d7a483a3dff" (UID: "1eea9d32-4e2a-4ad9-a402-7d7a483a3dff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.267935 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.267986 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.268001 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.268013 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxxm8\" (UniqueName: \"kubernetes.io/projected/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-kube-api-access-vxxm8\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.323109 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.589553 4948 generic.go:334] "Generic (PLEG): container finished" podID="1eea9d32-4e2a-4ad9-a402-7d7a483a3dff" containerID="3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911" exitCode=0 Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.589584 4948 generic.go:334] "Generic (PLEG): container finished" podID="1eea9d32-4e2a-4ad9-a402-7d7a483a3dff" containerID="b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a" exitCode=143 Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.589617 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff","Type":"ContainerDied","Data":"3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911"} Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.589641 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff","Type":"ContainerDied","Data":"b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a"} Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.589652 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff","Type":"ContainerDied","Data":"4ce5c5ab612acce40e4a349276cab71f6a92423acc9a5a2deff1b1f12271b01e"} Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.589667 4948 scope.go:117] "RemoveContainer" containerID="3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.589747 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.699472 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.704551 4948 scope.go:117] "RemoveContainer" containerID="b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.711745 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.730307 4948 scope.go:117] "RemoveContainer" containerID="3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911" Feb 20 08:24:25 crc kubenswrapper[4948]: E0220 08:24:25.736782 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911\": container with ID starting with 3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911 not found: ID does not exist" containerID="3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.736836 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911"} err="failed to get container status \"3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911\": rpc error: code = NotFound desc = could not find container \"3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911\": container with ID starting with 3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911 not found: ID does not exist" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.736892 4948 scope.go:117] "RemoveContainer" containerID="b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.737735 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" path="/var/lib/kubelet/pods/0894350a-0df8-4c04-b7c7-fd8f93139959/volumes" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.738456 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1eea9d32-4e2a-4ad9-a402-7d7a483a3dff" path="/var/lib/kubelet/pods/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff/volumes" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.738966 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:24:25 crc kubenswrapper[4948]: E0220 08:24:25.739281 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1eea9d32-4e2a-4ad9-a402-7d7a483a3dff" containerName="nova-metadata-log" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.739295 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="1eea9d32-4e2a-4ad9-a402-7d7a483a3dff" containerName="nova-metadata-log" Feb 20 08:24:25 crc kubenswrapper[4948]: E0220 08:24:25.739326 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1eea9d32-4e2a-4ad9-a402-7d7a483a3dff" containerName="nova-metadata-metadata" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.739332 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="1eea9d32-4e2a-4ad9-a402-7d7a483a3dff" containerName="nova-metadata-metadata" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.739499 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="1eea9d32-4e2a-4ad9-a402-7d7a483a3dff" containerName="nova-metadata-metadata" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.739520 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="1eea9d32-4e2a-4ad9-a402-7d7a483a3dff" containerName="nova-metadata-log" Feb 20 08:24:25 crc kubenswrapper[4948]: E0220 08:24:25.740204 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a\": container with ID starting with b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a not found: ID does not exist" containerID="b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.740252 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a"} err="failed to get container status \"b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a\": rpc error: code = NotFound desc = could not find container \"b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a\": container with ID starting with b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a not found: ID does not exist" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.740274 4948 scope.go:117] "RemoveContainer" containerID="3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.741068 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.743873 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911"} err="failed to get container status \"3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911\": rpc error: code = NotFound desc = could not find container \"3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911\": container with ID starting with 3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911 not found: ID does not exist" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.743910 4948 scope.go:117] "RemoveContainer" containerID="b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.744280 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.744361 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.747326 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a"} err="failed to get container status \"b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a\": rpc error: code = NotFound desc = could not find container \"b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a\": container with ID starting with b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a not found: ID does not exist" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.749826 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.785186 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-config-data\") pod \"nova-metadata-0\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " pod="openstack/nova-metadata-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.785277 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " pod="openstack/nova-metadata-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.785295 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49e86920-0969-4e9f-9094-be4fcc5e6909-logs\") pod \"nova-metadata-0\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " pod="openstack/nova-metadata-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.785322 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85fgw\" (UniqueName: \"kubernetes.io/projected/49e86920-0969-4e9f-9094-be4fcc5e6909-kube-api-access-85fgw\") pod \"nova-metadata-0\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " pod="openstack/nova-metadata-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.785336 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.785380 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " pod="openstack/nova-metadata-0" Feb 20 08:24:25 crc kubenswrapper[4948]: W0220 08:24:25.798950 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee88c030_4657_480c_99a0_90ee4c2b247c.slice/crio-7a7f2ca4a85292e0cdfdd80686c701b75dbf3b7d4a7696caa3251ac73f926e4a WatchSource:0}: Error finding container 7a7f2ca4a85292e0cdfdd80686c701b75dbf3b7d4a7696caa3251ac73f926e4a: Status 404 returned error can't find the container with id 7a7f2ca4a85292e0cdfdd80686c701b75dbf3b7d4a7696caa3251ac73f926e4a Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.886784 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-config-data\") pod \"nova-metadata-0\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " pod="openstack/nova-metadata-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.886846 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " pod="openstack/nova-metadata-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.886870 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49e86920-0969-4e9f-9094-be4fcc5e6909-logs\") pod \"nova-metadata-0\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " pod="openstack/nova-metadata-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.886918 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85fgw\" (UniqueName: \"kubernetes.io/projected/49e86920-0969-4e9f-9094-be4fcc5e6909-kube-api-access-85fgw\") pod \"nova-metadata-0\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " pod="openstack/nova-metadata-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.886987 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " pod="openstack/nova-metadata-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.888941 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49e86920-0969-4e9f-9094-be4fcc5e6909-logs\") pod \"nova-metadata-0\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " pod="openstack/nova-metadata-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.891174 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " pod="openstack/nova-metadata-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.893151 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-config-data\") pod \"nova-metadata-0\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " pod="openstack/nova-metadata-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.900833 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " pod="openstack/nova-metadata-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.913498 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85fgw\" (UniqueName: \"kubernetes.io/projected/49e86920-0969-4e9f-9094-be4fcc5e6909-kube-api-access-85fgw\") pod \"nova-metadata-0\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " pod="openstack/nova-metadata-0" Feb 20 08:24:26 crc kubenswrapper[4948]: I0220 08:24:26.057781 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 08:24:26 crc kubenswrapper[4948]: W0220 08:24:26.510332 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49e86920_0969_4e9f_9094_be4fcc5e6909.slice/crio-74fd9bc2eba68f1449066340009ee4df0ea7d5ad9b1c182e3bbc30ecc61defa2 WatchSource:0}: Error finding container 74fd9bc2eba68f1449066340009ee4df0ea7d5ad9b1c182e3bbc30ecc61defa2: Status 404 returned error can't find the container with id 74fd9bc2eba68f1449066340009ee4df0ea7d5ad9b1c182e3bbc30ecc61defa2 Feb 20 08:24:26 crc kubenswrapper[4948]: I0220 08:24:26.511380 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:24:26 crc kubenswrapper[4948]: I0220 08:24:26.601163 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"49e86920-0969-4e9f-9094-be4fcc5e6909","Type":"ContainerStarted","Data":"74fd9bc2eba68f1449066340009ee4df0ea7d5ad9b1c182e3bbc30ecc61defa2"} Feb 20 08:24:26 crc kubenswrapper[4948]: I0220 08:24:26.602677 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee88c030-4657-480c-99a0-90ee4c2b247c","Type":"ContainerStarted","Data":"7a7f2ca4a85292e0cdfdd80686c701b75dbf3b7d4a7696caa3251ac73f926e4a"} Feb 20 08:24:27 crc kubenswrapper[4948]: I0220 08:24:27.609752 4948 generic.go:334] "Generic (PLEG): container finished" podID="21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b" containerID="9a9d495cbe6ddcf0246661f3edacc2b8afe21e8ef43eca82e9140a67f21f1c6d" exitCode=0 Feb 20 08:24:27 crc kubenswrapper[4948]: I0220 08:24:27.609986 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qnlrn" event={"ID":"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b","Type":"ContainerDied","Data":"9a9d495cbe6ddcf0246661f3edacc2b8afe21e8ef43eca82e9140a67f21f1c6d"} Feb 20 08:24:27 crc kubenswrapper[4948]: I0220 08:24:27.615712 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"49e86920-0969-4e9f-9094-be4fcc5e6909","Type":"ContainerStarted","Data":"32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249"} Feb 20 08:24:27 crc kubenswrapper[4948]: I0220 08:24:27.615826 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"49e86920-0969-4e9f-9094-be4fcc5e6909","Type":"ContainerStarted","Data":"f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705"} Feb 20 08:24:27 crc kubenswrapper[4948]: I0220 08:24:27.658398 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.658377367 podStartE2EDuration="2.658377367s" podCreationTimestamp="2026-02-20 08:24:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:24:27.643721904 +0000 UTC m=+1116.618216724" watchObservedRunningTime="2026-02-20 08:24:27.658377367 +0000 UTC m=+1116.632872187" Feb 20 08:24:28 crc kubenswrapper[4948]: I0220 08:24:28.626794 4948 generic.go:334] "Generic (PLEG): container finished" podID="93f8e582-bd6a-44d4-a203-ba9950efcada" containerID="67cb9d3659cf07e0efb010d75eb62a0dddad421b940140c438f0a0714676a1ed" exitCode=0 Feb 20 08:24:28 crc kubenswrapper[4948]: I0220 08:24:28.626883 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-jl8ck" event={"ID":"93f8e582-bd6a-44d4-a203-ba9950efcada","Type":"ContainerDied","Data":"67cb9d3659cf07e0efb010d75eb62a0dddad421b940140c438f0a0714676a1ed"} Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.085641 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qnlrn" Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.186275 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-combined-ca-bundle\") pod \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\" (UID: \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\") " Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.186412 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2j4xb\" (UniqueName: \"kubernetes.io/projected/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-kube-api-access-2j4xb\") pod \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\" (UID: \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\") " Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.186556 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-scripts\") pod \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\" (UID: \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\") " Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.186603 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-config-data\") pod \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\" (UID: \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\") " Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.192390 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-scripts" (OuterVolumeSpecName: "scripts") pod "21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b" (UID: "21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.193049 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-kube-api-access-2j4xb" (OuterVolumeSpecName: "kube-api-access-2j4xb") pod "21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b" (UID: "21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b"). InnerVolumeSpecName "kube-api-access-2j4xb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.218159 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b" (UID: "21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.234133 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-config-data" (OuterVolumeSpecName: "config-data") pod "21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b" (UID: "21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.289359 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.290006 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.290048 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.290139 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2j4xb\" (UniqueName: \"kubernetes.io/projected/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-kube-api-access-2j4xb\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.641379 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qnlrn" Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.641761 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qnlrn" event={"ID":"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b","Type":"ContainerDied","Data":"12945171d0e3aefdd46fee90ddbf80a6cc5d07d1053f4e8bc8f96de548e35fcc"} Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.641857 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12945171d0e3aefdd46fee90ddbf80a6cc5d07d1053f4e8bc8f96de548e35fcc" Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.796401 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.796652 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="025de664-c84b-4bb4-b953-13d95e9eee5d" containerName="nova-api-log" containerID="cri-o://2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56" gracePeriod=30 Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.797118 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="025de664-c84b-4bb4-b953-13d95e9eee5d" containerName="nova-api-api" containerID="cri-o://a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7" gracePeriod=30 Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.816037 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.816252 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="60006453-1770-44da-bd53-89701da2fa43" containerName="nova-scheduler-scheduler" containerID="cri-o://789d33d79de6243b08689af1b9cf996bdaa6473c8e9a5246cde2871757156c6a" gracePeriod=30 Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.826955 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.827177 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="49e86920-0969-4e9f-9094-be4fcc5e6909" containerName="nova-metadata-log" containerID="cri-o://f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705" gracePeriod=30 Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.827310 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="49e86920-0969-4e9f-9094-be4fcc5e6909" containerName="nova-metadata-metadata" containerID="cri-o://32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249" gracePeriod=30 Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.163727 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-jl8ck" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.262168 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.314962 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-scripts\") pod \"93f8e582-bd6a-44d4-a203-ba9950efcada\" (UID: \"93f8e582-bd6a-44d4-a203-ba9950efcada\") " Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.315029 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-config-data\") pod \"93f8e582-bd6a-44d4-a203-ba9950efcada\" (UID: \"93f8e582-bd6a-44d4-a203-ba9950efcada\") " Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.315265 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czd6x\" (UniqueName: \"kubernetes.io/projected/93f8e582-bd6a-44d4-a203-ba9950efcada-kube-api-access-czd6x\") pod \"93f8e582-bd6a-44d4-a203-ba9950efcada\" (UID: \"93f8e582-bd6a-44d4-a203-ba9950efcada\") " Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.315296 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-combined-ca-bundle\") pod \"93f8e582-bd6a-44d4-a203-ba9950efcada\" (UID: \"93f8e582-bd6a-44d4-a203-ba9950efcada\") " Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.324859 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93f8e582-bd6a-44d4-a203-ba9950efcada-kube-api-access-czd6x" (OuterVolumeSpecName: "kube-api-access-czd6x") pod "93f8e582-bd6a-44d4-a203-ba9950efcada" (UID: "93f8e582-bd6a-44d4-a203-ba9950efcada"). InnerVolumeSpecName "kube-api-access-czd6x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.332102 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-scripts" (OuterVolumeSpecName: "scripts") pod "93f8e582-bd6a-44d4-a203-ba9950efcada" (UID: "93f8e582-bd6a-44d4-a203-ba9950efcada"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.378812 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-5dgw7"] Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.379071 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" podUID="63124eef-9456-44d1-a8ba-cc8297af3e9b" containerName="dnsmasq-dns" containerID="cri-o://724a55a11670e45fe3d2ce7a82832ba87e341cd787eee57c186dd30931fb8402" gracePeriod=10 Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.395401 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-config-data" (OuterVolumeSpecName: "config-data") pod "93f8e582-bd6a-44d4-a203-ba9950efcada" (UID: "93f8e582-bd6a-44d4-a203-ba9950efcada"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.401054 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "93f8e582-bd6a-44d4-a203-ba9950efcada" (UID: "93f8e582-bd6a-44d4-a203-ba9950efcada"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.418430 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.418455 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czd6x\" (UniqueName: \"kubernetes.io/projected/93f8e582-bd6a-44d4-a203-ba9950efcada-kube-api-access-czd6x\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.418472 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.418482 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.544095 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.549209 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.621409 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/025de664-c84b-4bb4-b953-13d95e9eee5d-combined-ca-bundle\") pod \"025de664-c84b-4bb4-b953-13d95e9eee5d\" (UID: \"025de664-c84b-4bb4-b953-13d95e9eee5d\") " Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.621480 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-nova-metadata-tls-certs\") pod \"49e86920-0969-4e9f-9094-be4fcc5e6909\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.621564 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85fgw\" (UniqueName: \"kubernetes.io/projected/49e86920-0969-4e9f-9094-be4fcc5e6909-kube-api-access-85fgw\") pod \"49e86920-0969-4e9f-9094-be4fcc5e6909\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.621621 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-config-data\") pod \"49e86920-0969-4e9f-9094-be4fcc5e6909\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.621743 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/025de664-c84b-4bb4-b953-13d95e9eee5d-config-data\") pod \"025de664-c84b-4bb4-b953-13d95e9eee5d\" (UID: \"025de664-c84b-4bb4-b953-13d95e9eee5d\") " Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.621791 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49e86920-0969-4e9f-9094-be4fcc5e6909-logs\") pod \"49e86920-0969-4e9f-9094-be4fcc5e6909\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.621822 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/025de664-c84b-4bb4-b953-13d95e9eee5d-logs\") pod \"025de664-c84b-4bb4-b953-13d95e9eee5d\" (UID: \"025de664-c84b-4bb4-b953-13d95e9eee5d\") " Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.621869 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blbc4\" (UniqueName: \"kubernetes.io/projected/025de664-c84b-4bb4-b953-13d95e9eee5d-kube-api-access-blbc4\") pod \"025de664-c84b-4bb4-b953-13d95e9eee5d\" (UID: \"025de664-c84b-4bb4-b953-13d95e9eee5d\") " Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.621920 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-combined-ca-bundle\") pod \"49e86920-0969-4e9f-9094-be4fcc5e6909\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.626413 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/025de664-c84b-4bb4-b953-13d95e9eee5d-logs" (OuterVolumeSpecName: "logs") pod "025de664-c84b-4bb4-b953-13d95e9eee5d" (UID: "025de664-c84b-4bb4-b953-13d95e9eee5d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.626724 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49e86920-0969-4e9f-9094-be4fcc5e6909-logs" (OuterVolumeSpecName: "logs") pod "49e86920-0969-4e9f-9094-be4fcc5e6909" (UID: "49e86920-0969-4e9f-9094-be4fcc5e6909"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.632560 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49e86920-0969-4e9f-9094-be4fcc5e6909-kube-api-access-85fgw" (OuterVolumeSpecName: "kube-api-access-85fgw") pod "49e86920-0969-4e9f-9094-be4fcc5e6909" (UID: "49e86920-0969-4e9f-9094-be4fcc5e6909"). InnerVolumeSpecName "kube-api-access-85fgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.642182 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/025de664-c84b-4bb4-b953-13d95e9eee5d-kube-api-access-blbc4" (OuterVolumeSpecName: "kube-api-access-blbc4") pod "025de664-c84b-4bb4-b953-13d95e9eee5d" (UID: "025de664-c84b-4bb4-b953-13d95e9eee5d"). InnerVolumeSpecName "kube-api-access-blbc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.651506 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee88c030-4657-480c-99a0-90ee4c2b247c","Type":"ContainerStarted","Data":"1630e56297ae1cd7ec343d1b51485df834ac4ef7ce2a756bf5a88dac19b42159"} Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.653507 4948 generic.go:334] "Generic (PLEG): container finished" podID="025de664-c84b-4bb4-b953-13d95e9eee5d" containerID="a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7" exitCode=0 Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.653670 4948 generic.go:334] "Generic (PLEG): container finished" podID="025de664-c84b-4bb4-b953-13d95e9eee5d" containerID="2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56" exitCode=143 Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.653571 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"025de664-c84b-4bb4-b953-13d95e9eee5d","Type":"ContainerDied","Data":"a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7"} Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.653571 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.654099 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"025de664-c84b-4bb4-b953-13d95e9eee5d","Type":"ContainerDied","Data":"2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56"} Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.654121 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"025de664-c84b-4bb4-b953-13d95e9eee5d","Type":"ContainerDied","Data":"a2ceacf4091ad26a00570ff4d1bfb8fcf3b35d5c7789387f917c60bdba3eb6b8"} Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.654138 4948 scope.go:117] "RemoveContainer" containerID="a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.666328 4948 generic.go:334] "Generic (PLEG): container finished" podID="63124eef-9456-44d1-a8ba-cc8297af3e9b" containerID="724a55a11670e45fe3d2ce7a82832ba87e341cd787eee57c186dd30931fb8402" exitCode=0 Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.666431 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" event={"ID":"63124eef-9456-44d1-a8ba-cc8297af3e9b","Type":"ContainerDied","Data":"724a55a11670e45fe3d2ce7a82832ba87e341cd787eee57c186dd30931fb8402"} Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.726687 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49e86920-0969-4e9f-9094-be4fcc5e6909-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.726713 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/025de664-c84b-4bb4-b953-13d95e9eee5d-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.726722 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blbc4\" (UniqueName: \"kubernetes.io/projected/025de664-c84b-4bb4-b953-13d95e9eee5d-kube-api-access-blbc4\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.726730 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85fgw\" (UniqueName: \"kubernetes.io/projected/49e86920-0969-4e9f-9094-be4fcc5e6909-kube-api-access-85fgw\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.742432 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/025de664-c84b-4bb4-b953-13d95e9eee5d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "025de664-c84b-4bb4-b953-13d95e9eee5d" (UID: "025de664-c84b-4bb4-b953-13d95e9eee5d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.742722 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-jl8ck" event={"ID":"93f8e582-bd6a-44d4-a203-ba9950efcada","Type":"ContainerDied","Data":"bc36e2f68570b6936d5e5dbd27c44aba36cbc45ceca2d3d6a770bf8b08d1d326"} Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.742750 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc36e2f68570b6936d5e5dbd27c44aba36cbc45ceca2d3d6a770bf8b08d1d326" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.742833 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-jl8ck" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.768654 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-config-data" (OuterVolumeSpecName: "config-data") pod "49e86920-0969-4e9f-9094-be4fcc5e6909" (UID: "49e86920-0969-4e9f-9094-be4fcc5e6909"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.778412 4948 scope.go:117] "RemoveContainer" containerID="2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.801185 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "49e86920-0969-4e9f-9094-be4fcc5e6909" (UID: "49e86920-0969-4e9f-9094-be4fcc5e6909"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.805109 4948 generic.go:334] "Generic (PLEG): container finished" podID="49e86920-0969-4e9f-9094-be4fcc5e6909" containerID="32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249" exitCode=0 Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.805140 4948 generic.go:334] "Generic (PLEG): container finished" podID="49e86920-0969-4e9f-9094-be4fcc5e6909" containerID="f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705" exitCode=143 Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.805161 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"49e86920-0969-4e9f-9094-be4fcc5e6909","Type":"ContainerDied","Data":"32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249"} Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.805192 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"49e86920-0969-4e9f-9094-be4fcc5e6909","Type":"ContainerDied","Data":"f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705"} Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.805203 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"49e86920-0969-4e9f-9094-be4fcc5e6909","Type":"ContainerDied","Data":"74fd9bc2eba68f1449066340009ee4df0ea7d5ad9b1c182e3bbc30ecc61defa2"} Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.805261 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.816335 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "49e86920-0969-4e9f-9094-be4fcc5e6909" (UID: "49e86920-0969-4e9f-9094-be4fcc5e6909"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.825681 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 20 08:24:30 crc kubenswrapper[4948]: E0220 08:24:30.826284 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93f8e582-bd6a-44d4-a203-ba9950efcada" containerName="nova-cell1-conductor-db-sync" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.826296 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="93f8e582-bd6a-44d4-a203-ba9950efcada" containerName="nova-cell1-conductor-db-sync" Feb 20 08:24:30 crc kubenswrapper[4948]: E0220 08:24:30.826308 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b" containerName="nova-manage" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.826314 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b" containerName="nova-manage" Feb 20 08:24:30 crc kubenswrapper[4948]: E0220 08:24:30.826332 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49e86920-0969-4e9f-9094-be4fcc5e6909" containerName="nova-metadata-metadata" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.826338 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="49e86920-0969-4e9f-9094-be4fcc5e6909" containerName="nova-metadata-metadata" Feb 20 08:24:30 crc kubenswrapper[4948]: E0220 08:24:30.826357 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="025de664-c84b-4bb4-b953-13d95e9eee5d" containerName="nova-api-api" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.826363 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="025de664-c84b-4bb4-b953-13d95e9eee5d" containerName="nova-api-api" Feb 20 08:24:30 crc kubenswrapper[4948]: E0220 08:24:30.826377 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49e86920-0969-4e9f-9094-be4fcc5e6909" containerName="nova-metadata-log" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.826383 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="49e86920-0969-4e9f-9094-be4fcc5e6909" containerName="nova-metadata-log" Feb 20 08:24:30 crc kubenswrapper[4948]: E0220 08:24:30.826392 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="025de664-c84b-4bb4-b953-13d95e9eee5d" containerName="nova-api-log" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.826398 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="025de664-c84b-4bb4-b953-13d95e9eee5d" containerName="nova-api-log" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.826554 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b" containerName="nova-manage" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.826566 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="49e86920-0969-4e9f-9094-be4fcc5e6909" containerName="nova-metadata-metadata" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.826579 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="025de664-c84b-4bb4-b953-13d95e9eee5d" containerName="nova-api-log" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.826587 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="025de664-c84b-4bb4-b953-13d95e9eee5d" containerName="nova-api-api" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.826595 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="49e86920-0969-4e9f-9094-be4fcc5e6909" containerName="nova-metadata-log" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.826604 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="93f8e582-bd6a-44d4-a203-ba9950efcada" containerName="nova-cell1-conductor-db-sync" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.827215 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.830337 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/025de664-c84b-4bb4-b953-13d95e9eee5d-config-data" (OuterVolumeSpecName: "config-data") pod "025de664-c84b-4bb4-b953-13d95e9eee5d" (UID: "025de664-c84b-4bb4-b953-13d95e9eee5d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.832174 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.832195 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/025de664-c84b-4bb4-b953-13d95e9eee5d-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.832207 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.832217 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/025de664-c84b-4bb4-b953-13d95e9eee5d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.832228 4948 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.832423 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.834599 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.874233 4948 scope.go:117] "RemoveContainer" containerID="a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7" Feb 20 08:24:30 crc kubenswrapper[4948]: E0220 08:24:30.876394 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7\": container with ID starting with a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7 not found: ID does not exist" containerID="a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.876434 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7"} err="failed to get container status \"a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7\": rpc error: code = NotFound desc = could not find container \"a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7\": container with ID starting with a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7 not found: ID does not exist" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.876458 4948 scope.go:117] "RemoveContainer" containerID="2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56" Feb 20 08:24:30 crc kubenswrapper[4948]: E0220 08:24:30.876799 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56\": container with ID starting with 2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56 not found: ID does not exist" containerID="2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.876814 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56"} err="failed to get container status \"2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56\": rpc error: code = NotFound desc = could not find container \"2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56\": container with ID starting with 2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56 not found: ID does not exist" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.876825 4948 scope.go:117] "RemoveContainer" containerID="a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.877034 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7"} err="failed to get container status \"a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7\": rpc error: code = NotFound desc = could not find container \"a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7\": container with ID starting with a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7 not found: ID does not exist" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.877047 4948 scope.go:117] "RemoveContainer" containerID="2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.877204 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56"} err="failed to get container status \"2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56\": rpc error: code = NotFound desc = could not find container \"2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56\": container with ID starting with 2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56 not found: ID does not exist" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.877217 4948 scope.go:117] "RemoveContainer" containerID="32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.934105 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e801ada9-0247-4b29-b262-04637e1f8452-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"e801ada9-0247-4b29-b262-04637e1f8452\") " pod="openstack/nova-cell1-conductor-0" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.934234 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e801ada9-0247-4b29-b262-04637e1f8452-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"e801ada9-0247-4b29-b262-04637e1f8452\") " pod="openstack/nova-cell1-conductor-0" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.934284 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dl7b\" (UniqueName: \"kubernetes.io/projected/e801ada9-0247-4b29-b262-04637e1f8452-kube-api-access-2dl7b\") pod \"nova-cell1-conductor-0\" (UID: \"e801ada9-0247-4b29-b262-04637e1f8452\") " pod="openstack/nova-cell1-conductor-0" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.948077 4948 scope.go:117] "RemoveContainer" containerID="f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.024496 4948 scope.go:117] "RemoveContainer" containerID="32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249" Feb 20 08:24:31 crc kubenswrapper[4948]: E0220 08:24:31.025456 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249\": container with ID starting with 32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249 not found: ID does not exist" containerID="32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.025497 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249"} err="failed to get container status \"32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249\": rpc error: code = NotFound desc = could not find container \"32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249\": container with ID starting with 32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249 not found: ID does not exist" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.025525 4948 scope.go:117] "RemoveContainer" containerID="f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705" Feb 20 08:24:31 crc kubenswrapper[4948]: E0220 08:24:31.025955 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705\": container with ID starting with f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705 not found: ID does not exist" containerID="f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.025993 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705"} err="failed to get container status \"f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705\": rpc error: code = NotFound desc = could not find container \"f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705\": container with ID starting with f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705 not found: ID does not exist" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.026006 4948 scope.go:117] "RemoveContainer" containerID="32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.026814 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249"} err="failed to get container status \"32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249\": rpc error: code = NotFound desc = could not find container \"32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249\": container with ID starting with 32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249 not found: ID does not exist" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.026835 4948 scope.go:117] "RemoveContainer" containerID="f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.027016 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705"} err="failed to get container status \"f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705\": rpc error: code = NotFound desc = could not find container \"f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705\": container with ID starting with f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705 not found: ID does not exist" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.035992 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dl7b\" (UniqueName: \"kubernetes.io/projected/e801ada9-0247-4b29-b262-04637e1f8452-kube-api-access-2dl7b\") pod \"nova-cell1-conductor-0\" (UID: \"e801ada9-0247-4b29-b262-04637e1f8452\") " pod="openstack/nova-cell1-conductor-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.036475 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e801ada9-0247-4b29-b262-04637e1f8452-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"e801ada9-0247-4b29-b262-04637e1f8452\") " pod="openstack/nova-cell1-conductor-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.036554 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e801ada9-0247-4b29-b262-04637e1f8452-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"e801ada9-0247-4b29-b262-04637e1f8452\") " pod="openstack/nova-cell1-conductor-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.042207 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e801ada9-0247-4b29-b262-04637e1f8452-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"e801ada9-0247-4b29-b262-04637e1f8452\") " pod="openstack/nova-cell1-conductor-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.055458 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.059715 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e801ada9-0247-4b29-b262-04637e1f8452-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"e801ada9-0247-4b29-b262-04637e1f8452\") " pod="openstack/nova-cell1-conductor-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.063957 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.066050 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dl7b\" (UniqueName: \"kubernetes.io/projected/e801ada9-0247-4b29-b262-04637e1f8452-kube-api-access-2dl7b\") pod \"nova-cell1-conductor-0\" (UID: \"e801ada9-0247-4b29-b262-04637e1f8452\") " pod="openstack/nova-cell1-conductor-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.072510 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.074099 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.076495 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.083684 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.136884 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.138722 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b950a35-d978-4338-88a6-ec96c9af048e-logs\") pod \"nova-api-0\" (UID: \"4b950a35-d978-4338-88a6-ec96c9af048e\") " pod="openstack/nova-api-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.138890 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b950a35-d978-4338-88a6-ec96c9af048e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4b950a35-d978-4338-88a6-ec96c9af048e\") " pod="openstack/nova-api-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.139220 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b950a35-d978-4338-88a6-ec96c9af048e-config-data\") pod \"nova-api-0\" (UID: \"4b950a35-d978-4338-88a6-ec96c9af048e\") " pod="openstack/nova-api-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.139281 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96ggh\" (UniqueName: \"kubernetes.io/projected/4b950a35-d978-4338-88a6-ec96c9af048e-kube-api-access-96ggh\") pod \"nova-api-0\" (UID: \"4b950a35-d978-4338-88a6-ec96c9af048e\") " pod="openstack/nova-api-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.151485 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.163208 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.177328 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:24:31 crc kubenswrapper[4948]: E0220 08:24:31.177805 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63124eef-9456-44d1-a8ba-cc8297af3e9b" containerName="dnsmasq-dns" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.177826 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="63124eef-9456-44d1-a8ba-cc8297af3e9b" containerName="dnsmasq-dns" Feb 20 08:24:31 crc kubenswrapper[4948]: E0220 08:24:31.177839 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63124eef-9456-44d1-a8ba-cc8297af3e9b" containerName="init" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.177847 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="63124eef-9456-44d1-a8ba-cc8297af3e9b" containerName="init" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.178104 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="63124eef-9456-44d1-a8ba-cc8297af3e9b" containerName="dnsmasq-dns" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.179118 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.181649 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.181876 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.199673 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.241390 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-ovsdbserver-sb\") pod \"63124eef-9456-44d1-a8ba-cc8297af3e9b\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.241781 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-ovsdbserver-nb\") pod \"63124eef-9456-44d1-a8ba-cc8297af3e9b\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.241812 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2cgj\" (UniqueName: \"kubernetes.io/projected/63124eef-9456-44d1-a8ba-cc8297af3e9b-kube-api-access-l2cgj\") pod \"63124eef-9456-44d1-a8ba-cc8297af3e9b\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.241866 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-dns-swift-storage-0\") pod \"63124eef-9456-44d1-a8ba-cc8297af3e9b\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.241938 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-config\") pod \"63124eef-9456-44d1-a8ba-cc8297af3e9b\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.242043 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-dns-svc\") pod \"63124eef-9456-44d1-a8ba-cc8297af3e9b\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.242326 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b950a35-d978-4338-88a6-ec96c9af048e-logs\") pod \"nova-api-0\" (UID: \"4b950a35-d978-4338-88a6-ec96c9af048e\") " pod="openstack/nova-api-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.242351 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7ch5\" (UniqueName: \"kubernetes.io/projected/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-kube-api-access-p7ch5\") pod \"nova-metadata-0\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " pod="openstack/nova-metadata-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.242398 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " pod="openstack/nova-metadata-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.242416 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b950a35-d978-4338-88a6-ec96c9af048e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4b950a35-d978-4338-88a6-ec96c9af048e\") " pod="openstack/nova-api-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.242435 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-config-data\") pod \"nova-metadata-0\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " pod="openstack/nova-metadata-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.242822 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b950a35-d978-4338-88a6-ec96c9af048e-logs\") pod \"nova-api-0\" (UID: \"4b950a35-d978-4338-88a6-ec96c9af048e\") " pod="openstack/nova-api-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.243880 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.244462 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " pod="openstack/nova-metadata-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.244492 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-logs\") pod \"nova-metadata-0\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " pod="openstack/nova-metadata-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.244553 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b950a35-d978-4338-88a6-ec96c9af048e-config-data\") pod \"nova-api-0\" (UID: \"4b950a35-d978-4338-88a6-ec96c9af048e\") " pod="openstack/nova-api-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.244572 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96ggh\" (UniqueName: \"kubernetes.io/projected/4b950a35-d978-4338-88a6-ec96c9af048e-kube-api-access-96ggh\") pod \"nova-api-0\" (UID: \"4b950a35-d978-4338-88a6-ec96c9af048e\") " pod="openstack/nova-api-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.252223 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63124eef-9456-44d1-a8ba-cc8297af3e9b-kube-api-access-l2cgj" (OuterVolumeSpecName: "kube-api-access-l2cgj") pod "63124eef-9456-44d1-a8ba-cc8297af3e9b" (UID: "63124eef-9456-44d1-a8ba-cc8297af3e9b"). InnerVolumeSpecName "kube-api-access-l2cgj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.253808 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b950a35-d978-4338-88a6-ec96c9af048e-config-data\") pod \"nova-api-0\" (UID: \"4b950a35-d978-4338-88a6-ec96c9af048e\") " pod="openstack/nova-api-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.262071 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b950a35-d978-4338-88a6-ec96c9af048e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4b950a35-d978-4338-88a6-ec96c9af048e\") " pod="openstack/nova-api-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.265283 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96ggh\" (UniqueName: \"kubernetes.io/projected/4b950a35-d978-4338-88a6-ec96c9af048e-kube-api-access-96ggh\") pod \"nova-api-0\" (UID: \"4b950a35-d978-4338-88a6-ec96c9af048e\") " pod="openstack/nova-api-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.317936 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-config" (OuterVolumeSpecName: "config") pod "63124eef-9456-44d1-a8ba-cc8297af3e9b" (UID: "63124eef-9456-44d1-a8ba-cc8297af3e9b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.317683 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "63124eef-9456-44d1-a8ba-cc8297af3e9b" (UID: "63124eef-9456-44d1-a8ba-cc8297af3e9b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.325316 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "63124eef-9456-44d1-a8ba-cc8297af3e9b" (UID: "63124eef-9456-44d1-a8ba-cc8297af3e9b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.345719 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "63124eef-9456-44d1-a8ba-cc8297af3e9b" (UID: "63124eef-9456-44d1-a8ba-cc8297af3e9b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.345974 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " pod="openstack/nova-metadata-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.346032 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-logs\") pod \"nova-metadata-0\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " pod="openstack/nova-metadata-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.346151 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7ch5\" (UniqueName: \"kubernetes.io/projected/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-kube-api-access-p7ch5\") pod \"nova-metadata-0\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " pod="openstack/nova-metadata-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.346198 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " pod="openstack/nova-metadata-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.346260 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-config-data\") pod \"nova-metadata-0\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " pod="openstack/nova-metadata-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.346357 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "63124eef-9456-44d1-a8ba-cc8297af3e9b" (UID: "63124eef-9456-44d1-a8ba-cc8297af3e9b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.346365 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.346431 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2cgj\" (UniqueName: \"kubernetes.io/projected/63124eef-9456-44d1-a8ba-cc8297af3e9b-kube-api-access-l2cgj\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.346494 4948 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.346506 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.346515 4948 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.347387 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-logs\") pod \"nova-metadata-0\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " pod="openstack/nova-metadata-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.351915 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " pod="openstack/nova-metadata-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.354162 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " pod="openstack/nova-metadata-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.354402 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-config-data\") pod \"nova-metadata-0\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " pod="openstack/nova-metadata-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.364837 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7ch5\" (UniqueName: \"kubernetes.io/projected/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-kube-api-access-p7ch5\") pod \"nova-metadata-0\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " pod="openstack/nova-metadata-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.448531 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.448708 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.524770 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.571407 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.653676 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dl2cc\" (UniqueName: \"kubernetes.io/projected/60006453-1770-44da-bd53-89701da2fa43-kube-api-access-dl2cc\") pod \"60006453-1770-44da-bd53-89701da2fa43\" (UID: \"60006453-1770-44da-bd53-89701da2fa43\") " Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.653726 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60006453-1770-44da-bd53-89701da2fa43-combined-ca-bundle\") pod \"60006453-1770-44da-bd53-89701da2fa43\" (UID: \"60006453-1770-44da-bd53-89701da2fa43\") " Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.653858 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60006453-1770-44da-bd53-89701da2fa43-config-data\") pod \"60006453-1770-44da-bd53-89701da2fa43\" (UID: \"60006453-1770-44da-bd53-89701da2fa43\") " Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.660390 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60006453-1770-44da-bd53-89701da2fa43-kube-api-access-dl2cc" (OuterVolumeSpecName: "kube-api-access-dl2cc") pod "60006453-1770-44da-bd53-89701da2fa43" (UID: "60006453-1770-44da-bd53-89701da2fa43"). InnerVolumeSpecName "kube-api-access-dl2cc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.681497 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60006453-1770-44da-bd53-89701da2fa43-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "60006453-1770-44da-bd53-89701da2fa43" (UID: "60006453-1770-44da-bd53-89701da2fa43"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.688976 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60006453-1770-44da-bd53-89701da2fa43-config-data" (OuterVolumeSpecName: "config-data") pod "60006453-1770-44da-bd53-89701da2fa43" (UID: "60006453-1770-44da-bd53-89701da2fa43"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.761593 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dl2cc\" (UniqueName: \"kubernetes.io/projected/60006453-1770-44da-bd53-89701da2fa43-kube-api-access-dl2cc\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.761897 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60006453-1770-44da-bd53-89701da2fa43-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.761915 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60006453-1770-44da-bd53-89701da2fa43-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.768290 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="025de664-c84b-4bb4-b953-13d95e9eee5d" path="/var/lib/kubelet/pods/025de664-c84b-4bb4-b953-13d95e9eee5d/volumes" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.769453 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49e86920-0969-4e9f-9094-be4fcc5e6909" path="/var/lib/kubelet/pods/49e86920-0969-4e9f-9094-be4fcc5e6909/volumes" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.771595 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.844374 4948 generic.go:334] "Generic (PLEG): container finished" podID="60006453-1770-44da-bd53-89701da2fa43" containerID="789d33d79de6243b08689af1b9cf996bdaa6473c8e9a5246cde2871757156c6a" exitCode=0 Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.844540 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.844638 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"60006453-1770-44da-bd53-89701da2fa43","Type":"ContainerDied","Data":"789d33d79de6243b08689af1b9cf996bdaa6473c8e9a5246cde2871757156c6a"} Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.844697 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"60006453-1770-44da-bd53-89701da2fa43","Type":"ContainerDied","Data":"22ac2dae6a489cbbc45577f9a64760c4218b1a60e308183ad1c3246e8923567d"} Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.844747 4948 scope.go:117] "RemoveContainer" containerID="789d33d79de6243b08689af1b9cf996bdaa6473c8e9a5246cde2871757156c6a" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.855787 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee88c030-4657-480c-99a0-90ee4c2b247c","Type":"ContainerStarted","Data":"1bcad95c714fce375a1b0a71dbbd9424565d0d8570fd34397817043b9bfe2e88"} Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.858750 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" event={"ID":"63124eef-9456-44d1-a8ba-cc8297af3e9b","Type":"ContainerDied","Data":"b02b398d101ad6f54b2deb0825184d401d1e5b0ad4e38257f4120a442c5a5509"} Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.858840 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.860025 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"e801ada9-0247-4b29-b262-04637e1f8452","Type":"ContainerStarted","Data":"4b659e360d8bb9cff850961f25c15821b1a120efd9472c53adb045fc0c3cbcca"} Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.877419 4948 scope.go:117] "RemoveContainer" containerID="789d33d79de6243b08689af1b9cf996bdaa6473c8e9a5246cde2871757156c6a" Feb 20 08:24:31 crc kubenswrapper[4948]: E0220 08:24:31.878478 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"789d33d79de6243b08689af1b9cf996bdaa6473c8e9a5246cde2871757156c6a\": container with ID starting with 789d33d79de6243b08689af1b9cf996bdaa6473c8e9a5246cde2871757156c6a not found: ID does not exist" containerID="789d33d79de6243b08689af1b9cf996bdaa6473c8e9a5246cde2871757156c6a" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.878554 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"789d33d79de6243b08689af1b9cf996bdaa6473c8e9a5246cde2871757156c6a"} err="failed to get container status \"789d33d79de6243b08689af1b9cf996bdaa6473c8e9a5246cde2871757156c6a\": rpc error: code = NotFound desc = could not find container \"789d33d79de6243b08689af1b9cf996bdaa6473c8e9a5246cde2871757156c6a\": container with ID starting with 789d33d79de6243b08689af1b9cf996bdaa6473c8e9a5246cde2871757156c6a not found: ID does not exist" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.878609 4948 scope.go:117] "RemoveContainer" containerID="724a55a11670e45fe3d2ce7a82832ba87e341cd787eee57c186dd30931fb8402" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.918676 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.931646 4948 scope.go:117] "RemoveContainer" containerID="56e628ea61223b6bdd0b21738d110aaaa403a55ddd583614f61bc57a6cdb16fd" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.940932 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.953815 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-5dgw7"] Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.972327 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-5dgw7"] Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.981039 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 08:24:31 crc kubenswrapper[4948]: E0220 08:24:31.981729 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60006453-1770-44da-bd53-89701da2fa43" containerName="nova-scheduler-scheduler" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.981766 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="60006453-1770-44da-bd53-89701da2fa43" containerName="nova-scheduler-scheduler" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.982060 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="60006453-1770-44da-bd53-89701da2fa43" containerName="nova-scheduler-scheduler" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.984522 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.987172 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.991480 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.031183 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:24:32 crc kubenswrapper[4948]: W0220 08:24:32.045800 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod659ea49b_7e8f_4c95_ac69_ec743d0f95ed.slice/crio-aaa4c4ee0564a915596564697d18943937054d1b42224076a88c2793a4afd7d0 WatchSource:0}: Error finding container aaa4c4ee0564a915596564697d18943937054d1b42224076a88c2793a4afd7d0: Status 404 returned error can't find the container with id aaa4c4ee0564a915596564697d18943937054d1b42224076a88c2793a4afd7d0 Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.069100 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a76350cb-da45-403e-8934-ef53b257ea27-config-data\") pod \"nova-scheduler-0\" (UID: \"a76350cb-da45-403e-8934-ef53b257ea27\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.069567 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cb4nn\" (UniqueName: \"kubernetes.io/projected/a76350cb-da45-403e-8934-ef53b257ea27-kube-api-access-cb4nn\") pod \"nova-scheduler-0\" (UID: \"a76350cb-da45-403e-8934-ef53b257ea27\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.069692 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a76350cb-da45-403e-8934-ef53b257ea27-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a76350cb-da45-403e-8934-ef53b257ea27\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:32 crc kubenswrapper[4948]: W0220 08:24:32.107152 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b950a35_d978_4338_88a6_ec96c9af048e.slice/crio-d66a33a2d7dbdb39f5f061898cb4a5149bc962e88f36c5fc69dae56ca6bdd17a WatchSource:0}: Error finding container d66a33a2d7dbdb39f5f061898cb4a5149bc962e88f36c5fc69dae56ca6bdd17a: Status 404 returned error can't find the container with id d66a33a2d7dbdb39f5f061898cb4a5149bc962e88f36c5fc69dae56ca6bdd17a Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.108360 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.171023 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a76350cb-da45-403e-8934-ef53b257ea27-config-data\") pod \"nova-scheduler-0\" (UID: \"a76350cb-da45-403e-8934-ef53b257ea27\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.171056 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cb4nn\" (UniqueName: \"kubernetes.io/projected/a76350cb-da45-403e-8934-ef53b257ea27-kube-api-access-cb4nn\") pod \"nova-scheduler-0\" (UID: \"a76350cb-da45-403e-8934-ef53b257ea27\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.171130 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a76350cb-da45-403e-8934-ef53b257ea27-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a76350cb-da45-403e-8934-ef53b257ea27\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.179025 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a76350cb-da45-403e-8934-ef53b257ea27-config-data\") pod \"nova-scheduler-0\" (UID: \"a76350cb-da45-403e-8934-ef53b257ea27\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.179542 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a76350cb-da45-403e-8934-ef53b257ea27-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a76350cb-da45-403e-8934-ef53b257ea27\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.186863 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cb4nn\" (UniqueName: \"kubernetes.io/projected/a76350cb-da45-403e-8934-ef53b257ea27-kube-api-access-cb4nn\") pod \"nova-scheduler-0\" (UID: \"a76350cb-da45-403e-8934-ef53b257ea27\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.273094 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.695186 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 08:24:32 crc kubenswrapper[4948]: W0220 08:24:32.697709 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda76350cb_da45_403e_8934_ef53b257ea27.slice/crio-6fbbcc32e0efcd622515f0c849bc2e7cae40139cc47cc3aeb9187b65bfe28ca5 WatchSource:0}: Error finding container 6fbbcc32e0efcd622515f0c849bc2e7cae40139cc47cc3aeb9187b65bfe28ca5: Status 404 returned error can't find the container with id 6fbbcc32e0efcd622515f0c849bc2e7cae40139cc47cc3aeb9187b65bfe28ca5 Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.869121 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a76350cb-da45-403e-8934-ef53b257ea27","Type":"ContainerStarted","Data":"6fbbcc32e0efcd622515f0c849bc2e7cae40139cc47cc3aeb9187b65bfe28ca5"} Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.871849 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"e801ada9-0247-4b29-b262-04637e1f8452","Type":"ContainerStarted","Data":"681c4bfe91487bec6fbff0de13c9517419b14290b69f54162d9b1bbc3572c765"} Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.872497 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.876868 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b950a35-d978-4338-88a6-ec96c9af048e","Type":"ContainerStarted","Data":"6ef713a538c9adc323b40a21aaf7370bc2c92d050478ddd9bfe5e00490dfc4a2"} Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.876897 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b950a35-d978-4338-88a6-ec96c9af048e","Type":"ContainerStarted","Data":"d0e4271ae345ef60d4741e0350d3336cff2d587f6e410c83f608d1092728d8fa"} Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.876908 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b950a35-d978-4338-88a6-ec96c9af048e","Type":"ContainerStarted","Data":"d66a33a2d7dbdb39f5f061898cb4a5149bc962e88f36c5fc69dae56ca6bdd17a"} Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.879399 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"659ea49b-7e8f-4c95-ac69-ec743d0f95ed","Type":"ContainerStarted","Data":"22196e9ec797dbe697aeeef9301c64792cbe1db93384c86a9e5389ca3fe652c9"} Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.879428 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"659ea49b-7e8f-4c95-ac69-ec743d0f95ed","Type":"ContainerStarted","Data":"63251b467317ee4b7966810a99bfea1f41b91c2ea1703852741d551304c8bdaf"} Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.879441 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"659ea49b-7e8f-4c95-ac69-ec743d0f95ed","Type":"ContainerStarted","Data":"aaa4c4ee0564a915596564697d18943937054d1b42224076a88c2793a4afd7d0"} Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.884323 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee88c030-4657-480c-99a0-90ee4c2b247c","Type":"ContainerStarted","Data":"0574f1ab13d44d1f98f65e5fc2fb1727a44b35bb855f4f5336cf1da833d95b3a"} Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.899070 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.899046777 podStartE2EDuration="2.899046777s" podCreationTimestamp="2026-02-20 08:24:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:24:32.88760045 +0000 UTC m=+1121.862095270" watchObservedRunningTime="2026-02-20 08:24:32.899046777 +0000 UTC m=+1121.873541617" Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.914698 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.914677324 podStartE2EDuration="1.914677324s" podCreationTimestamp="2026-02-20 08:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:24:32.908593937 +0000 UTC m=+1121.883088787" watchObservedRunningTime="2026-02-20 08:24:32.914677324 +0000 UTC m=+1121.889172154" Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.929641 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.929622974 podStartE2EDuration="1.929622974s" podCreationTimestamp="2026-02-20 08:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:24:32.928411695 +0000 UTC m=+1121.902906535" watchObservedRunningTime="2026-02-20 08:24:32.929622974 +0000 UTC m=+1121.904117794" Feb 20 08:24:33 crc kubenswrapper[4948]: I0220 08:24:33.749814 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60006453-1770-44da-bd53-89701da2fa43" path="/var/lib/kubelet/pods/60006453-1770-44da-bd53-89701da2fa43/volumes" Feb 20 08:24:33 crc kubenswrapper[4948]: I0220 08:24:33.753170 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63124eef-9456-44d1-a8ba-cc8297af3e9b" path="/var/lib/kubelet/pods/63124eef-9456-44d1-a8ba-cc8297af3e9b/volumes" Feb 20 08:24:33 crc kubenswrapper[4948]: I0220 08:24:33.897114 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a76350cb-da45-403e-8934-ef53b257ea27","Type":"ContainerStarted","Data":"7fdf44e32062d9617c830886caa9e40ee3055d0b238b4f4ec53507c904d02787"} Feb 20 08:24:33 crc kubenswrapper[4948]: I0220 08:24:33.931444 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.9314236 podStartE2EDuration="2.9314236s" podCreationTimestamp="2026-02-20 08:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:24:33.915587688 +0000 UTC m=+1122.890082508" watchObservedRunningTime="2026-02-20 08:24:33.9314236 +0000 UTC m=+1122.905918420" Feb 20 08:24:34 crc kubenswrapper[4948]: I0220 08:24:34.911754 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee88c030-4657-480c-99a0-90ee4c2b247c","Type":"ContainerStarted","Data":"430724be7dcee7dfd82861aa2d26928ec4bed184137610d057128b4eb3d23e14"} Feb 20 08:24:34 crc kubenswrapper[4948]: I0220 08:24:34.951847 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.677508246 podStartE2EDuration="10.951818044s" podCreationTimestamp="2026-02-20 08:24:24 +0000 UTC" firstStartedPulling="2026-02-20 08:24:25.800995295 +0000 UTC m=+1114.775490135" lastFinishedPulling="2026-02-20 08:24:34.075305103 +0000 UTC m=+1123.049799933" observedRunningTime="2026-02-20 08:24:34.941510935 +0000 UTC m=+1123.916005795" watchObservedRunningTime="2026-02-20 08:24:34.951818044 +0000 UTC m=+1123.926312904" Feb 20 08:24:35 crc kubenswrapper[4948]: I0220 08:24:35.926465 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 20 08:24:36 crc kubenswrapper[4948]: I0220 08:24:36.282966 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Feb 20 08:24:36 crc kubenswrapper[4948]: I0220 08:24:36.525739 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 20 08:24:36 crc kubenswrapper[4948]: I0220 08:24:36.525848 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 20 08:24:37 crc kubenswrapper[4948]: I0220 08:24:37.273845 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 20 08:24:38 crc kubenswrapper[4948]: I0220 08:24:38.025362 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:24:38 crc kubenswrapper[4948]: I0220 08:24:38.025748 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:24:38 crc kubenswrapper[4948]: I0220 08:24:38.025804 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:24:38 crc kubenswrapper[4948]: I0220 08:24:38.026438 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8f0b8a806330c630468951158113ee0ecaad5dbddb81b09c063752a17f21282c"} pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 08:24:38 crc kubenswrapper[4948]: I0220 08:24:38.026490 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" containerID="cri-o://8f0b8a806330c630468951158113ee0ecaad5dbddb81b09c063752a17f21282c" gracePeriod=600 Feb 20 08:24:38 crc kubenswrapper[4948]: I0220 08:24:38.967676 4948 generic.go:334] "Generic (PLEG): container finished" podID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerID="8f0b8a806330c630468951158113ee0ecaad5dbddb81b09c063752a17f21282c" exitCode=0 Feb 20 08:24:38 crc kubenswrapper[4948]: I0220 08:24:38.967802 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerDied","Data":"8f0b8a806330c630468951158113ee0ecaad5dbddb81b09c063752a17f21282c"} Feb 20 08:24:38 crc kubenswrapper[4948]: I0220 08:24:38.968226 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"d9efd6716995fcdcebdcf01e76fc7927735e34b1da4918444f90e59bd0f6d815"} Feb 20 08:24:38 crc kubenswrapper[4948]: I0220 08:24:38.968275 4948 scope.go:117] "RemoveContainer" containerID="df8e056c06ac241c46a927c4e3f5a28f39d042bf13a8c672499d528786dbbbb0" Feb 20 08:24:41 crc kubenswrapper[4948]: I0220 08:24:41.449787 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 20 08:24:41 crc kubenswrapper[4948]: I0220 08:24:41.450463 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 20 08:24:41 crc kubenswrapper[4948]: I0220 08:24:41.525393 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 20 08:24:41 crc kubenswrapper[4948]: I0220 08:24:41.525454 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 20 08:24:42 crc kubenswrapper[4948]: I0220 08:24:42.274283 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 20 08:24:42 crc kubenswrapper[4948]: I0220 08:24:42.304732 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 20 08:24:42 crc kubenswrapper[4948]: I0220 08:24:42.532694 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4b950a35-d978-4338-88a6-ec96c9af048e" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.197:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 20 08:24:42 crc kubenswrapper[4948]: I0220 08:24:42.533442 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4b950a35-d978-4338-88a6-ec96c9af048e" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.197:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 20 08:24:42 crc kubenswrapper[4948]: I0220 08:24:42.560309 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="659ea49b-7e8f-4c95-ac69-ec743d0f95ed" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.198:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 20 08:24:42 crc kubenswrapper[4948]: I0220 08:24:42.560510 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="659ea49b-7e8f-4c95-ac69-ec743d0f95ed" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.198:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 20 08:24:43 crc kubenswrapper[4948]: I0220 08:24:43.064452 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.453726 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.455931 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.456580 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.456749 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.464291 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.467104 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.533770 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.535720 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.540152 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.692155 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-t56xb"] Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.693547 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.739238 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-t56xb"] Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.788022 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-config\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.788127 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.788174 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.788235 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-db22s\" (UniqueName: \"kubernetes.io/projected/c1847849-cdf4-4ea6-a540-c01503a003b2-kube-api-access-db22s\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.788269 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.788303 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.890027 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.890145 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-db22s\" (UniqueName: \"kubernetes.io/projected/c1847849-cdf4-4ea6-a540-c01503a003b2-kube-api-access-db22s\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.890184 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.890225 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.890321 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-config\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.890369 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.890907 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.891376 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.891542 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-config\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.892040 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.892118 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.911195 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-db22s\" (UniqueName: \"kubernetes.io/projected/c1847849-cdf4-4ea6-a540-c01503a003b2-kube-api-access-db22s\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:52 crc kubenswrapper[4948]: I0220 08:24:52.023413 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:52 crc kubenswrapper[4948]: I0220 08:24:52.146725 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 20 08:24:52 crc kubenswrapper[4948]: I0220 08:24:52.570848 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-t56xb"] Feb 20 08:24:52 crc kubenswrapper[4948]: W0220 08:24:52.596817 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1847849_cdf4_4ea6_a540_c01503a003b2.slice/crio-4b70d018879f9b501eed3beab4a8c0781964989ca45e86dd4327f865ae78f6bf WatchSource:0}: Error finding container 4b70d018879f9b501eed3beab4a8c0781964989ca45e86dd4327f865ae78f6bf: Status 404 returned error can't find the container with id 4b70d018879f9b501eed3beab4a8c0781964989ca45e86dd4327f865ae78f6bf Feb 20 08:24:53 crc kubenswrapper[4948]: I0220 08:24:53.148894 4948 generic.go:334] "Generic (PLEG): container finished" podID="c1847849-cdf4-4ea6-a540-c01503a003b2" containerID="6b70c340bd4972b75cf8aa94cef8eb5d658d1c6345ae3959074d416c3683e47a" exitCode=0 Feb 20 08:24:53 crc kubenswrapper[4948]: I0220 08:24:53.149035 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" event={"ID":"c1847849-cdf4-4ea6-a540-c01503a003b2","Type":"ContainerDied","Data":"6b70c340bd4972b75cf8aa94cef8eb5d658d1c6345ae3959074d416c3683e47a"} Feb 20 08:24:53 crc kubenswrapper[4948]: I0220 08:24:53.149527 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" event={"ID":"c1847849-cdf4-4ea6-a540-c01503a003b2","Type":"ContainerStarted","Data":"4b70d018879f9b501eed3beab4a8c0781964989ca45e86dd4327f865ae78f6bf"} Feb 20 08:24:53 crc kubenswrapper[4948]: I0220 08:24:53.669343 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:24:53 crc kubenswrapper[4948]: I0220 08:24:53.669632 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="ceilometer-central-agent" containerID="cri-o://1630e56297ae1cd7ec343d1b51485df834ac4ef7ce2a756bf5a88dac19b42159" gracePeriod=30 Feb 20 08:24:53 crc kubenswrapper[4948]: I0220 08:24:53.669746 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="ceilometer-notification-agent" containerID="cri-o://1bcad95c714fce375a1b0a71dbbd9424565d0d8570fd34397817043b9bfe2e88" gracePeriod=30 Feb 20 08:24:53 crc kubenswrapper[4948]: I0220 08:24:53.669768 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="proxy-httpd" containerID="cri-o://430724be7dcee7dfd82861aa2d26928ec4bed184137610d057128b4eb3d23e14" gracePeriod=30 Feb 20 08:24:53 crc kubenswrapper[4948]: I0220 08:24:53.669822 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="sg-core" containerID="cri-o://0574f1ab13d44d1f98f65e5fc2fb1727a44b35bb855f4f5336cf1da833d95b3a" gracePeriod=30 Feb 20 08:24:53 crc kubenswrapper[4948]: I0220 08:24:53.701748 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Feb 20 08:24:53 crc kubenswrapper[4948]: I0220 08:24:53.971387 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 20 08:24:54 crc kubenswrapper[4948]: I0220 08:24:54.163603 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" event={"ID":"c1847849-cdf4-4ea6-a540-c01503a003b2","Type":"ContainerStarted","Data":"5c65f781c2289e22368c8df4db43181d77d0c56ecf63f298c154f0b8af27a14f"} Feb 20 08:24:54 crc kubenswrapper[4948]: I0220 08:24:54.163933 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:54 crc kubenswrapper[4948]: I0220 08:24:54.166802 4948 generic.go:334] "Generic (PLEG): container finished" podID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerID="430724be7dcee7dfd82861aa2d26928ec4bed184137610d057128b4eb3d23e14" exitCode=0 Feb 20 08:24:54 crc kubenswrapper[4948]: I0220 08:24:54.166825 4948 generic.go:334] "Generic (PLEG): container finished" podID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerID="0574f1ab13d44d1f98f65e5fc2fb1727a44b35bb855f4f5336cf1da833d95b3a" exitCode=2 Feb 20 08:24:54 crc kubenswrapper[4948]: I0220 08:24:54.166833 4948 generic.go:334] "Generic (PLEG): container finished" podID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerID="1630e56297ae1cd7ec343d1b51485df834ac4ef7ce2a756bf5a88dac19b42159" exitCode=0 Feb 20 08:24:54 crc kubenswrapper[4948]: I0220 08:24:54.166919 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee88c030-4657-480c-99a0-90ee4c2b247c","Type":"ContainerDied","Data":"430724be7dcee7dfd82861aa2d26928ec4bed184137610d057128b4eb3d23e14"} Feb 20 08:24:54 crc kubenswrapper[4948]: I0220 08:24:54.167024 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee88c030-4657-480c-99a0-90ee4c2b247c","Type":"ContainerDied","Data":"0574f1ab13d44d1f98f65e5fc2fb1727a44b35bb855f4f5336cf1da833d95b3a"} Feb 20 08:24:54 crc kubenswrapper[4948]: I0220 08:24:54.167046 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee88c030-4657-480c-99a0-90ee4c2b247c","Type":"ContainerDied","Data":"1630e56297ae1cd7ec343d1b51485df834ac4ef7ce2a756bf5a88dac19b42159"} Feb 20 08:24:54 crc kubenswrapper[4948]: I0220 08:24:54.167289 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4b950a35-d978-4338-88a6-ec96c9af048e" containerName="nova-api-log" containerID="cri-o://d0e4271ae345ef60d4741e0350d3336cff2d587f6e410c83f608d1092728d8fa" gracePeriod=30 Feb 20 08:24:54 crc kubenswrapper[4948]: I0220 08:24:54.167328 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4b950a35-d978-4338-88a6-ec96c9af048e" containerName="nova-api-api" containerID="cri-o://6ef713a538c9adc323b40a21aaf7370bc2c92d050478ddd9bfe5e00490dfc4a2" gracePeriod=30 Feb 20 08:24:54 crc kubenswrapper[4948]: I0220 08:24:54.196656 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" podStartSLOduration=3.196639074 podStartE2EDuration="3.196639074s" podCreationTimestamp="2026-02-20 08:24:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:24:54.194924773 +0000 UTC m=+1143.169419593" watchObservedRunningTime="2026-02-20 08:24:54.196639074 +0000 UTC m=+1143.171133894" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.034384 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.170178 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-combined-ca-bundle\") pod \"ad615fe7-b01a-4e4c-85b1-e028e5cdd866\" (UID: \"ad615fe7-b01a-4e4c-85b1-e028e5cdd866\") " Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.170313 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-config-data\") pod \"ad615fe7-b01a-4e4c-85b1-e028e5cdd866\" (UID: \"ad615fe7-b01a-4e4c-85b1-e028e5cdd866\") " Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.170360 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ll5q\" (UniqueName: \"kubernetes.io/projected/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-kube-api-access-2ll5q\") pod \"ad615fe7-b01a-4e4c-85b1-e028e5cdd866\" (UID: \"ad615fe7-b01a-4e4c-85b1-e028e5cdd866\") " Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.187328 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-kube-api-access-2ll5q" (OuterVolumeSpecName: "kube-api-access-2ll5q") pod "ad615fe7-b01a-4e4c-85b1-e028e5cdd866" (UID: "ad615fe7-b01a-4e4c-85b1-e028e5cdd866"). InnerVolumeSpecName "kube-api-access-2ll5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.190965 4948 generic.go:334] "Generic (PLEG): container finished" podID="4b950a35-d978-4338-88a6-ec96c9af048e" containerID="d0e4271ae345ef60d4741e0350d3336cff2d587f6e410c83f608d1092728d8fa" exitCode=143 Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.191136 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b950a35-d978-4338-88a6-ec96c9af048e","Type":"ContainerDied","Data":"d0e4271ae345ef60d4741e0350d3336cff2d587f6e410c83f608d1092728d8fa"} Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.192556 4948 generic.go:334] "Generic (PLEG): container finished" podID="ad615fe7-b01a-4e4c-85b1-e028e5cdd866" containerID="2242b79213417225ca0b9ded36fc0195241950291c46926c9dff4400bf7ca4e9" exitCode=137 Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.193653 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ad615fe7-b01a-4e4c-85b1-e028e5cdd866","Type":"ContainerDied","Data":"2242b79213417225ca0b9ded36fc0195241950291c46926c9dff4400bf7ca4e9"} Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.193703 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ad615fe7-b01a-4e4c-85b1-e028e5cdd866","Type":"ContainerDied","Data":"2e2fb5966e87b5f5075085440ad9aaba0868d3b137c6af57f293094a97faf543"} Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.193717 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.193722 4948 scope.go:117] "RemoveContainer" containerID="2242b79213417225ca0b9ded36fc0195241950291c46926c9dff4400bf7ca4e9" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.232366 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ad615fe7-b01a-4e4c-85b1-e028e5cdd866" (UID: "ad615fe7-b01a-4e4c-85b1-e028e5cdd866"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.252340 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-config-data" (OuterVolumeSpecName: "config-data") pod "ad615fe7-b01a-4e4c-85b1-e028e5cdd866" (UID: "ad615fe7-b01a-4e4c-85b1-e028e5cdd866"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.257484 4948 scope.go:117] "RemoveContainer" containerID="2242b79213417225ca0b9ded36fc0195241950291c46926c9dff4400bf7ca4e9" Feb 20 08:24:55 crc kubenswrapper[4948]: E0220 08:24:55.257830 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2242b79213417225ca0b9ded36fc0195241950291c46926c9dff4400bf7ca4e9\": container with ID starting with 2242b79213417225ca0b9ded36fc0195241950291c46926c9dff4400bf7ca4e9 not found: ID does not exist" containerID="2242b79213417225ca0b9ded36fc0195241950291c46926c9dff4400bf7ca4e9" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.257860 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2242b79213417225ca0b9ded36fc0195241950291c46926c9dff4400bf7ca4e9"} err="failed to get container status \"2242b79213417225ca0b9ded36fc0195241950291c46926c9dff4400bf7ca4e9\": rpc error: code = NotFound desc = could not find container \"2242b79213417225ca0b9ded36fc0195241950291c46926c9dff4400bf7ca4e9\": container with ID starting with 2242b79213417225ca0b9ded36fc0195241950291c46926c9dff4400bf7ca4e9 not found: ID does not exist" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.274448 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.274514 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.274530 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ll5q\" (UniqueName: \"kubernetes.io/projected/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-kube-api-access-2ll5q\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.324447 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.194:3000/\": dial tcp 10.217.0.194:3000: connect: connection refused" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.528702 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.536938 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.552131 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 20 08:24:55 crc kubenswrapper[4948]: E0220 08:24:55.553087 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad615fe7-b01a-4e4c-85b1-e028e5cdd866" containerName="nova-cell1-novncproxy-novncproxy" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.553134 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad615fe7-b01a-4e4c-85b1-e028e5cdd866" containerName="nova-cell1-novncproxy-novncproxy" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.553485 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad615fe7-b01a-4e4c-85b1-e028e5cdd866" containerName="nova-cell1-novncproxy-novncproxy" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.554151 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.558057 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.558185 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.558426 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.569476 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.682126 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c00f24b-fedb-4747-91c4-3d0b551a7288-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c00f24b-fedb-4747-91c4-3d0b551a7288\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.682195 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c00f24b-fedb-4747-91c4-3d0b551a7288-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c00f24b-fedb-4747-91c4-3d0b551a7288\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.682333 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c00f24b-fedb-4747-91c4-3d0b551a7288-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c00f24b-fedb-4747-91c4-3d0b551a7288\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.682426 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6tdc\" (UniqueName: \"kubernetes.io/projected/9c00f24b-fedb-4747-91c4-3d0b551a7288-kube-api-access-n6tdc\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c00f24b-fedb-4747-91c4-3d0b551a7288\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.682463 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c00f24b-fedb-4747-91c4-3d0b551a7288-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c00f24b-fedb-4747-91c4-3d0b551a7288\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.734413 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad615fe7-b01a-4e4c-85b1-e028e5cdd866" path="/var/lib/kubelet/pods/ad615fe7-b01a-4e4c-85b1-e028e5cdd866/volumes" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.784639 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c00f24b-fedb-4747-91c4-3d0b551a7288-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c00f24b-fedb-4747-91c4-3d0b551a7288\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.784716 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c00f24b-fedb-4747-91c4-3d0b551a7288-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c00f24b-fedb-4747-91c4-3d0b551a7288\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.784856 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c00f24b-fedb-4747-91c4-3d0b551a7288-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c00f24b-fedb-4747-91c4-3d0b551a7288\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.784991 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6tdc\" (UniqueName: \"kubernetes.io/projected/9c00f24b-fedb-4747-91c4-3d0b551a7288-kube-api-access-n6tdc\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c00f24b-fedb-4747-91c4-3d0b551a7288\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.785035 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c00f24b-fedb-4747-91c4-3d0b551a7288-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c00f24b-fedb-4747-91c4-3d0b551a7288\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.789503 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c00f24b-fedb-4747-91c4-3d0b551a7288-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c00f24b-fedb-4747-91c4-3d0b551a7288\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.790215 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c00f24b-fedb-4747-91c4-3d0b551a7288-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c00f24b-fedb-4747-91c4-3d0b551a7288\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.792409 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c00f24b-fedb-4747-91c4-3d0b551a7288-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c00f24b-fedb-4747-91c4-3d0b551a7288\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.792797 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c00f24b-fedb-4747-91c4-3d0b551a7288-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c00f24b-fedb-4747-91c4-3d0b551a7288\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.804676 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6tdc\" (UniqueName: \"kubernetes.io/projected/9c00f24b-fedb-4747-91c4-3d0b551a7288-kube-api-access-n6tdc\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c00f24b-fedb-4747-91c4-3d0b551a7288\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.870682 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:56 crc kubenswrapper[4948]: I0220 08:24:56.344046 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 20 08:24:56 crc kubenswrapper[4948]: W0220 08:24:56.344830 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c00f24b_fedb_4747_91c4_3d0b551a7288.slice/crio-bffecdc61bc8fd8e92ede8f2c61eaf9d4b0abac0d945cb8a1261deb94cd475b9 WatchSource:0}: Error finding container bffecdc61bc8fd8e92ede8f2c61eaf9d4b0abac0d945cb8a1261deb94cd475b9: Status 404 returned error can't find the container with id bffecdc61bc8fd8e92ede8f2c61eaf9d4b0abac0d945cb8a1261deb94cd475b9 Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.238086 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9c00f24b-fedb-4747-91c4-3d0b551a7288","Type":"ContainerStarted","Data":"aed40c01186dd40a2c4f854e7c4f35be79dfb8528a952fc8f05c38778e541dff"} Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.238533 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9c00f24b-fedb-4747-91c4-3d0b551a7288","Type":"ContainerStarted","Data":"bffecdc61bc8fd8e92ede8f2c61eaf9d4b0abac0d945cb8a1261deb94cd475b9"} Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.242431 4948 generic.go:334] "Generic (PLEG): container finished" podID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerID="1bcad95c714fce375a1b0a71dbbd9424565d0d8570fd34397817043b9bfe2e88" exitCode=0 Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.242488 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee88c030-4657-480c-99a0-90ee4c2b247c","Type":"ContainerDied","Data":"1bcad95c714fce375a1b0a71dbbd9424565d0d8570fd34397817043b9bfe2e88"} Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.266100 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.266071646 podStartE2EDuration="2.266071646s" podCreationTimestamp="2026-02-20 08:24:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:24:57.255575583 +0000 UTC m=+1146.230070443" watchObservedRunningTime="2026-02-20 08:24:57.266071646 +0000 UTC m=+1146.240566476" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.521622 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.621895 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-config-data\") pod \"ee88c030-4657-480c-99a0-90ee4c2b247c\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.622008 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-combined-ca-bundle\") pod \"ee88c030-4657-480c-99a0-90ee4c2b247c\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.622077 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee88c030-4657-480c-99a0-90ee4c2b247c-run-httpd\") pod \"ee88c030-4657-480c-99a0-90ee4c2b247c\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.622101 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee88c030-4657-480c-99a0-90ee4c2b247c-log-httpd\") pod \"ee88c030-4657-480c-99a0-90ee4c2b247c\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.622157 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-scripts\") pod \"ee88c030-4657-480c-99a0-90ee4c2b247c\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.622206 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrzbr\" (UniqueName: \"kubernetes.io/projected/ee88c030-4657-480c-99a0-90ee4c2b247c-kube-api-access-rrzbr\") pod \"ee88c030-4657-480c-99a0-90ee4c2b247c\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.622311 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-sg-core-conf-yaml\") pod \"ee88c030-4657-480c-99a0-90ee4c2b247c\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.623014 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee88c030-4657-480c-99a0-90ee4c2b247c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ee88c030-4657-480c-99a0-90ee4c2b247c" (UID: "ee88c030-4657-480c-99a0-90ee4c2b247c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.626392 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee88c030-4657-480c-99a0-90ee4c2b247c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ee88c030-4657-480c-99a0-90ee4c2b247c" (UID: "ee88c030-4657-480c-99a0-90ee4c2b247c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.648931 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee88c030-4657-480c-99a0-90ee4c2b247c-kube-api-access-rrzbr" (OuterVolumeSpecName: "kube-api-access-rrzbr") pod "ee88c030-4657-480c-99a0-90ee4c2b247c" (UID: "ee88c030-4657-480c-99a0-90ee4c2b247c"). InnerVolumeSpecName "kube-api-access-rrzbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.662620 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-scripts" (OuterVolumeSpecName: "scripts") pod "ee88c030-4657-480c-99a0-90ee4c2b247c" (UID: "ee88c030-4657-480c-99a0-90ee4c2b247c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.682058 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ee88c030-4657-480c-99a0-90ee4c2b247c" (UID: "ee88c030-4657-480c-99a0-90ee4c2b247c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.721159 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.724416 4948 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee88c030-4657-480c-99a0-90ee4c2b247c-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.724449 4948 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee88c030-4657-480c-99a0-90ee4c2b247c-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.724460 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.724471 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrzbr\" (UniqueName: \"kubernetes.io/projected/ee88c030-4657-480c-99a0-90ee4c2b247c-kube-api-access-rrzbr\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.724483 4948 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.757108 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ee88c030-4657-480c-99a0-90ee4c2b247c" (UID: "ee88c030-4657-480c-99a0-90ee4c2b247c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.781184 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-config-data" (OuterVolumeSpecName: "config-data") pod "ee88c030-4657-480c-99a0-90ee4c2b247c" (UID: "ee88c030-4657-480c-99a0-90ee4c2b247c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.826684 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96ggh\" (UniqueName: \"kubernetes.io/projected/4b950a35-d978-4338-88a6-ec96c9af048e-kube-api-access-96ggh\") pod \"4b950a35-d978-4338-88a6-ec96c9af048e\" (UID: \"4b950a35-d978-4338-88a6-ec96c9af048e\") " Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.826908 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b950a35-d978-4338-88a6-ec96c9af048e-config-data\") pod \"4b950a35-d978-4338-88a6-ec96c9af048e\" (UID: \"4b950a35-d978-4338-88a6-ec96c9af048e\") " Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.826938 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b950a35-d978-4338-88a6-ec96c9af048e-logs\") pod \"4b950a35-d978-4338-88a6-ec96c9af048e\" (UID: \"4b950a35-d978-4338-88a6-ec96c9af048e\") " Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.827029 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b950a35-d978-4338-88a6-ec96c9af048e-combined-ca-bundle\") pod \"4b950a35-d978-4338-88a6-ec96c9af048e\" (UID: \"4b950a35-d978-4338-88a6-ec96c9af048e\") " Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.827642 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.827667 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.828520 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b950a35-d978-4338-88a6-ec96c9af048e-logs" (OuterVolumeSpecName: "logs") pod "4b950a35-d978-4338-88a6-ec96c9af048e" (UID: "4b950a35-d978-4338-88a6-ec96c9af048e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.834165 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b950a35-d978-4338-88a6-ec96c9af048e-kube-api-access-96ggh" (OuterVolumeSpecName: "kube-api-access-96ggh") pod "4b950a35-d978-4338-88a6-ec96c9af048e" (UID: "4b950a35-d978-4338-88a6-ec96c9af048e"). InnerVolumeSpecName "kube-api-access-96ggh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.862355 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b950a35-d978-4338-88a6-ec96c9af048e-config-data" (OuterVolumeSpecName: "config-data") pod "4b950a35-d978-4338-88a6-ec96c9af048e" (UID: "4b950a35-d978-4338-88a6-ec96c9af048e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.865900 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b950a35-d978-4338-88a6-ec96c9af048e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4b950a35-d978-4338-88a6-ec96c9af048e" (UID: "4b950a35-d978-4338-88a6-ec96c9af048e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.930086 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96ggh\" (UniqueName: \"kubernetes.io/projected/4b950a35-d978-4338-88a6-ec96c9af048e-kube-api-access-96ggh\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.930122 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b950a35-d978-4338-88a6-ec96c9af048e-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.930135 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b950a35-d978-4338-88a6-ec96c9af048e-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.930145 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b950a35-d978-4338-88a6-ec96c9af048e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.263958 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.264007 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee88c030-4657-480c-99a0-90ee4c2b247c","Type":"ContainerDied","Data":"7a7f2ca4a85292e0cdfdd80686c701b75dbf3b7d4a7696caa3251ac73f926e4a"} Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.265013 4948 scope.go:117] "RemoveContainer" containerID="430724be7dcee7dfd82861aa2d26928ec4bed184137610d057128b4eb3d23e14" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.269247 4948 generic.go:334] "Generic (PLEG): container finished" podID="4b950a35-d978-4338-88a6-ec96c9af048e" containerID="6ef713a538c9adc323b40a21aaf7370bc2c92d050478ddd9bfe5e00490dfc4a2" exitCode=0 Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.269303 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.269362 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b950a35-d978-4338-88a6-ec96c9af048e","Type":"ContainerDied","Data":"6ef713a538c9adc323b40a21aaf7370bc2c92d050478ddd9bfe5e00490dfc4a2"} Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.269422 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b950a35-d978-4338-88a6-ec96c9af048e","Type":"ContainerDied","Data":"d66a33a2d7dbdb39f5f061898cb4a5149bc962e88f36c5fc69dae56ca6bdd17a"} Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.326952 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.336206 4948 scope.go:117] "RemoveContainer" containerID="0574f1ab13d44d1f98f65e5fc2fb1727a44b35bb855f4f5336cf1da833d95b3a" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.355733 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.365868 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.385245 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.396098 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 20 08:24:58 crc kubenswrapper[4948]: E0220 08:24:58.400138 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b950a35-d978-4338-88a6-ec96c9af048e" containerName="nova-api-log" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.400181 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b950a35-d978-4338-88a6-ec96c9af048e" containerName="nova-api-log" Feb 20 08:24:58 crc kubenswrapper[4948]: E0220 08:24:58.400195 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="ceilometer-notification-agent" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.400202 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="ceilometer-notification-agent" Feb 20 08:24:58 crc kubenswrapper[4948]: E0220 08:24:58.400212 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b950a35-d978-4338-88a6-ec96c9af048e" containerName="nova-api-api" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.400218 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b950a35-d978-4338-88a6-ec96c9af048e" containerName="nova-api-api" Feb 20 08:24:58 crc kubenswrapper[4948]: E0220 08:24:58.400241 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="ceilometer-central-agent" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.400246 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="ceilometer-central-agent" Feb 20 08:24:58 crc kubenswrapper[4948]: E0220 08:24:58.400258 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="proxy-httpd" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.400264 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="proxy-httpd" Feb 20 08:24:58 crc kubenswrapper[4948]: E0220 08:24:58.400274 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="sg-core" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.400281 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="sg-core" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.400443 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b950a35-d978-4338-88a6-ec96c9af048e" containerName="nova-api-api" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.400454 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="ceilometer-central-agent" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.400470 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="sg-core" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.400484 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="ceilometer-notification-agent" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.400494 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b950a35-d978-4338-88a6-ec96c9af048e" containerName="nova-api-log" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.400506 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="proxy-httpd" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.400518 4948 scope.go:117] "RemoveContainer" containerID="1bcad95c714fce375a1b0a71dbbd9424565d0d8570fd34397817043b9bfe2e88" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.401424 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.403749 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.403871 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.403879 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.405955 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.414221 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.416693 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.426778 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.427127 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.427267 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.432509 4948 scope.go:117] "RemoveContainer" containerID="1630e56297ae1cd7ec343d1b51485df834ac4ef7ce2a756bf5a88dac19b42159" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.464283 4948 scope.go:117] "RemoveContainer" containerID="6ef713a538c9adc323b40a21aaf7370bc2c92d050478ddd9bfe5e00490dfc4a2" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.483954 4948 scope.go:117] "RemoveContainer" containerID="d0e4271ae345ef60d4741e0350d3336cff2d587f6e410c83f608d1092728d8fa" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.501096 4948 scope.go:117] "RemoveContainer" containerID="6ef713a538c9adc323b40a21aaf7370bc2c92d050478ddd9bfe5e00490dfc4a2" Feb 20 08:24:58 crc kubenswrapper[4948]: E0220 08:24:58.501457 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ef713a538c9adc323b40a21aaf7370bc2c92d050478ddd9bfe5e00490dfc4a2\": container with ID starting with 6ef713a538c9adc323b40a21aaf7370bc2c92d050478ddd9bfe5e00490dfc4a2 not found: ID does not exist" containerID="6ef713a538c9adc323b40a21aaf7370bc2c92d050478ddd9bfe5e00490dfc4a2" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.501504 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ef713a538c9adc323b40a21aaf7370bc2c92d050478ddd9bfe5e00490dfc4a2"} err="failed to get container status \"6ef713a538c9adc323b40a21aaf7370bc2c92d050478ddd9bfe5e00490dfc4a2\": rpc error: code = NotFound desc = could not find container \"6ef713a538c9adc323b40a21aaf7370bc2c92d050478ddd9bfe5e00490dfc4a2\": container with ID starting with 6ef713a538c9adc323b40a21aaf7370bc2c92d050478ddd9bfe5e00490dfc4a2 not found: ID does not exist" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.501535 4948 scope.go:117] "RemoveContainer" containerID="d0e4271ae345ef60d4741e0350d3336cff2d587f6e410c83f608d1092728d8fa" Feb 20 08:24:58 crc kubenswrapper[4948]: E0220 08:24:58.501832 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0e4271ae345ef60d4741e0350d3336cff2d587f6e410c83f608d1092728d8fa\": container with ID starting with d0e4271ae345ef60d4741e0350d3336cff2d587f6e410c83f608d1092728d8fa not found: ID does not exist" containerID="d0e4271ae345ef60d4741e0350d3336cff2d587f6e410c83f608d1092728d8fa" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.501855 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0e4271ae345ef60d4741e0350d3336cff2d587f6e410c83f608d1092728d8fa"} err="failed to get container status \"d0e4271ae345ef60d4741e0350d3336cff2d587f6e410c83f608d1092728d8fa\": rpc error: code = NotFound desc = could not find container \"d0e4271ae345ef60d4741e0350d3336cff2d587f6e410c83f608d1092728d8fa\": container with ID starting with d0e4271ae345ef60d4741e0350d3336cff2d587f6e410c83f608d1092728d8fa not found: ID does not exist" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.540912 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.541019 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61657c77-bb29-4241-9d88-5e13bccd3f4e-logs\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.541056 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-config-data\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.541082 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.541343 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.541418 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-scripts\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.541570 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-public-tls-certs\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.541624 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.541662 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-run-httpd\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.541682 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwwjr\" (UniqueName: \"kubernetes.io/projected/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-kube-api-access-mwwjr\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.541717 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zr8fh\" (UniqueName: \"kubernetes.io/projected/61657c77-bb29-4241-9d88-5e13bccd3f4e-kube-api-access-zr8fh\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.541735 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-config-data\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.541789 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-log-httpd\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.643807 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.643849 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-scripts\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.643904 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-public-tls-certs\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.643929 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.643947 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwwjr\" (UniqueName: \"kubernetes.io/projected/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-kube-api-access-mwwjr\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.643965 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-run-httpd\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.643997 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zr8fh\" (UniqueName: \"kubernetes.io/projected/61657c77-bb29-4241-9d88-5e13bccd3f4e-kube-api-access-zr8fh\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.644010 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-config-data\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.644034 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-log-httpd\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.644056 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.644500 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-run-httpd\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.644765 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61657c77-bb29-4241-9d88-5e13bccd3f4e-logs\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.644797 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-config-data\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.644795 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-log-httpd\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.644815 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.645284 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61657c77-bb29-4241-9d88-5e13bccd3f4e-logs\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.649386 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.649383 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-config-data\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.649590 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-public-tls-certs\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.650796 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.650984 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-config-data\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.654206 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-scripts\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.659767 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.665171 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.667124 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zr8fh\" (UniqueName: \"kubernetes.io/projected/61657c77-bb29-4241-9d88-5e13bccd3f4e-kube-api-access-zr8fh\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.667650 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwwjr\" (UniqueName: \"kubernetes.io/projected/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-kube-api-access-mwwjr\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.718024 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.743047 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:24:59 crc kubenswrapper[4948]: I0220 08:24:59.222463 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 20 08:24:59 crc kubenswrapper[4948]: W0220 08:24:59.227679 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod61657c77_bb29_4241_9d88_5e13bccd3f4e.slice/crio-cff172040120f907a742a7013b9d3e9198e3f80ccb81cd1545dba19de50f6adb WatchSource:0}: Error finding container cff172040120f907a742a7013b9d3e9198e3f80ccb81cd1545dba19de50f6adb: Status 404 returned error can't find the container with id cff172040120f907a742a7013b9d3e9198e3f80ccb81cd1545dba19de50f6adb Feb 20 08:24:59 crc kubenswrapper[4948]: W0220 08:24:59.231382 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbe0e09bc_edd8_4ea9_8a59_ab630ed91cd8.slice/crio-48f0f09932c5f951969d8a7a57693456f1c18049262597223bdbf2f58c0804b2 WatchSource:0}: Error finding container 48f0f09932c5f951969d8a7a57693456f1c18049262597223bdbf2f58c0804b2: Status 404 returned error can't find the container with id 48f0f09932c5f951969d8a7a57693456f1c18049262597223bdbf2f58c0804b2 Feb 20 08:24:59 crc kubenswrapper[4948]: I0220 08:24:59.234325 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:24:59 crc kubenswrapper[4948]: I0220 08:24:59.293026 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8","Type":"ContainerStarted","Data":"48f0f09932c5f951969d8a7a57693456f1c18049262597223bdbf2f58c0804b2"} Feb 20 08:24:59 crc kubenswrapper[4948]: I0220 08:24:59.296842 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"61657c77-bb29-4241-9d88-5e13bccd3f4e","Type":"ContainerStarted","Data":"cff172040120f907a742a7013b9d3e9198e3f80ccb81cd1545dba19de50f6adb"} Feb 20 08:24:59 crc kubenswrapper[4948]: I0220 08:24:59.739604 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b950a35-d978-4338-88a6-ec96c9af048e" path="/var/lib/kubelet/pods/4b950a35-d978-4338-88a6-ec96c9af048e/volumes" Feb 20 08:24:59 crc kubenswrapper[4948]: I0220 08:24:59.741074 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" path="/var/lib/kubelet/pods/ee88c030-4657-480c-99a0-90ee4c2b247c/volumes" Feb 20 08:25:00 crc kubenswrapper[4948]: I0220 08:25:00.312275 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8","Type":"ContainerStarted","Data":"d03b827c5401b1d3755870ce6ce51a7201888c77eda8df6a4497ae437d66e88a"} Feb 20 08:25:00 crc kubenswrapper[4948]: I0220 08:25:00.314794 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"61657c77-bb29-4241-9d88-5e13bccd3f4e","Type":"ContainerStarted","Data":"185d2ffc7312a2c6e5587955ab1e83610f6970f92e1e567908cad389309483bc"} Feb 20 08:25:00 crc kubenswrapper[4948]: I0220 08:25:00.314842 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"61657c77-bb29-4241-9d88-5e13bccd3f4e","Type":"ContainerStarted","Data":"38deb2b3c4145711eaa3d46abb6b2827da68608a05ee0a36a26658e34f7dc7af"} Feb 20 08:25:00 crc kubenswrapper[4948]: I0220 08:25:00.351775 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.351760001 podStartE2EDuration="2.351760001s" podCreationTimestamp="2026-02-20 08:24:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:25:00.340812587 +0000 UTC m=+1149.315307417" watchObservedRunningTime="2026-02-20 08:25:00.351760001 +0000 UTC m=+1149.326254821" Feb 20 08:25:00 crc kubenswrapper[4948]: I0220 08:25:00.871668 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:25:01 crc kubenswrapper[4948]: I0220 08:25:01.324918 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8","Type":"ContainerStarted","Data":"e04f1d97ea889929a68ccc1a2406ae3771d7b5ea9af14caaec52d943991404e0"} Feb 20 08:25:01 crc kubenswrapper[4948]: I0220 08:25:01.325328 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8","Type":"ContainerStarted","Data":"43e2a82447778e081804b3e50b6f485e50305ec4eb1bbe2f231f259d11b87542"} Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.026185 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.104027 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-x22px"] Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.104282 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-757b4f8459-x22px" podUID="538c1b6f-424d-48e3-a92b-7534fbf865fc" containerName="dnsmasq-dns" containerID="cri-o://3a9357bad2b3dca80dc83315104b35d29a1b980c7594ca0611bf4b94b36eb262" gracePeriod=10 Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.338038 4948 generic.go:334] "Generic (PLEG): container finished" podID="538c1b6f-424d-48e3-a92b-7534fbf865fc" containerID="3a9357bad2b3dca80dc83315104b35d29a1b980c7594ca0611bf4b94b36eb262" exitCode=0 Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.338091 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-x22px" event={"ID":"538c1b6f-424d-48e3-a92b-7534fbf865fc","Type":"ContainerDied","Data":"3a9357bad2b3dca80dc83315104b35d29a1b980c7594ca0611bf4b94b36eb262"} Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.637499 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.725439 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-ovsdbserver-nb\") pod \"538c1b6f-424d-48e3-a92b-7534fbf865fc\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.725588 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hh28w\" (UniqueName: \"kubernetes.io/projected/538c1b6f-424d-48e3-a92b-7534fbf865fc-kube-api-access-hh28w\") pod \"538c1b6f-424d-48e3-a92b-7534fbf865fc\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.725683 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-config\") pod \"538c1b6f-424d-48e3-a92b-7534fbf865fc\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.725720 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-dns-swift-storage-0\") pod \"538c1b6f-424d-48e3-a92b-7534fbf865fc\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.725783 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-ovsdbserver-sb\") pod \"538c1b6f-424d-48e3-a92b-7534fbf865fc\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.725807 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-dns-svc\") pod \"538c1b6f-424d-48e3-a92b-7534fbf865fc\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.734221 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/538c1b6f-424d-48e3-a92b-7534fbf865fc-kube-api-access-hh28w" (OuterVolumeSpecName: "kube-api-access-hh28w") pod "538c1b6f-424d-48e3-a92b-7534fbf865fc" (UID: "538c1b6f-424d-48e3-a92b-7534fbf865fc"). InnerVolumeSpecName "kube-api-access-hh28w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.778035 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "538c1b6f-424d-48e3-a92b-7534fbf865fc" (UID: "538c1b6f-424d-48e3-a92b-7534fbf865fc"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.780815 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "538c1b6f-424d-48e3-a92b-7534fbf865fc" (UID: "538c1b6f-424d-48e3-a92b-7534fbf865fc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.784027 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "538c1b6f-424d-48e3-a92b-7534fbf865fc" (UID: "538c1b6f-424d-48e3-a92b-7534fbf865fc"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:25:02 crc kubenswrapper[4948]: E0220 08:25:02.799387 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-ovsdbserver-nb podName:538c1b6f-424d-48e3-a92b-7534fbf865fc nodeName:}" failed. No retries permitted until 2026-02-20 08:25:03.299359908 +0000 UTC m=+1152.273854728 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "ovsdbserver-nb" (UniqueName: "kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-ovsdbserver-nb") pod "538c1b6f-424d-48e3-a92b-7534fbf865fc" (UID: "538c1b6f-424d-48e3-a92b-7534fbf865fc") : error deleting /var/lib/kubelet/pods/538c1b6f-424d-48e3-a92b-7534fbf865fc/volume-subpaths: remove /var/lib/kubelet/pods/538c1b6f-424d-48e3-a92b-7534fbf865fc/volume-subpaths: no such file or directory Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.799670 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-config" (OuterVolumeSpecName: "config") pod "538c1b6f-424d-48e3-a92b-7534fbf865fc" (UID: "538c1b6f-424d-48e3-a92b-7534fbf865fc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.832857 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hh28w\" (UniqueName: \"kubernetes.io/projected/538c1b6f-424d-48e3-a92b-7534fbf865fc-kube-api-access-hh28w\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.832915 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.832929 4948 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.832943 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.832986 4948 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:03 crc kubenswrapper[4948]: I0220 08:25:03.342878 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-ovsdbserver-nb\") pod \"538c1b6f-424d-48e3-a92b-7534fbf865fc\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " Feb 20 08:25:03 crc kubenswrapper[4948]: I0220 08:25:03.345787 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "538c1b6f-424d-48e3-a92b-7534fbf865fc" (UID: "538c1b6f-424d-48e3-a92b-7534fbf865fc"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:25:03 crc kubenswrapper[4948]: I0220 08:25:03.358181 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-x22px" event={"ID":"538c1b6f-424d-48e3-a92b-7534fbf865fc","Type":"ContainerDied","Data":"ca84f60c8f59ce30814af7db2ac25fea0b071bd0c109ef4bdad869f451268b2b"} Feb 20 08:25:03 crc kubenswrapper[4948]: I0220 08:25:03.358196 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:25:03 crc kubenswrapper[4948]: I0220 08:25:03.358860 4948 scope.go:117] "RemoveContainer" containerID="3a9357bad2b3dca80dc83315104b35d29a1b980c7594ca0611bf4b94b36eb262" Feb 20 08:25:03 crc kubenswrapper[4948]: I0220 08:25:03.362536 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8","Type":"ContainerStarted","Data":"a1bbb73bd50190b5b08c60773344adf5c06efb89abf70ae7646fd73534b398b9"} Feb 20 08:25:03 crc kubenswrapper[4948]: I0220 08:25:03.363379 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 20 08:25:03 crc kubenswrapper[4948]: I0220 08:25:03.387716 4948 scope.go:117] "RemoveContainer" containerID="70074ef7c3184ccb90626f2f932107dda3ac2287a8b654203232efa428fd6294" Feb 20 08:25:03 crc kubenswrapper[4948]: I0220 08:25:03.421540 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.942159117 podStartE2EDuration="5.421504641s" podCreationTimestamp="2026-02-20 08:24:58 +0000 UTC" firstStartedPulling="2026-02-20 08:24:59.233563857 +0000 UTC m=+1148.208058677" lastFinishedPulling="2026-02-20 08:25:02.712909381 +0000 UTC m=+1151.687404201" observedRunningTime="2026-02-20 08:25:03.399280294 +0000 UTC m=+1152.373775114" watchObservedRunningTime="2026-02-20 08:25:03.421504641 +0000 UTC m=+1152.395999461" Feb 20 08:25:03 crc kubenswrapper[4948]: I0220 08:25:03.429362 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-x22px"] Feb 20 08:25:03 crc kubenswrapper[4948]: I0220 08:25:03.440134 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-x22px"] Feb 20 08:25:03 crc kubenswrapper[4948]: I0220 08:25:03.447058 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:03 crc kubenswrapper[4948]: I0220 08:25:03.733611 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="538c1b6f-424d-48e3-a92b-7534fbf865fc" path="/var/lib/kubelet/pods/538c1b6f-424d-48e3-a92b-7534fbf865fc/volumes" Feb 20 08:25:05 crc kubenswrapper[4948]: I0220 08:25:05.871490 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:25:05 crc kubenswrapper[4948]: I0220 08:25:05.893947 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.410893 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.608958 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-mg24f"] Feb 20 08:25:06 crc kubenswrapper[4948]: E0220 08:25:06.609462 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="538c1b6f-424d-48e3-a92b-7534fbf865fc" containerName="dnsmasq-dns" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.609477 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="538c1b6f-424d-48e3-a92b-7534fbf865fc" containerName="dnsmasq-dns" Feb 20 08:25:06 crc kubenswrapper[4948]: E0220 08:25:06.609499 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="538c1b6f-424d-48e3-a92b-7534fbf865fc" containerName="init" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.609507 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="538c1b6f-424d-48e3-a92b-7534fbf865fc" containerName="init" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.609737 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="538c1b6f-424d-48e3-a92b-7534fbf865fc" containerName="dnsmasq-dns" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.610514 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mg24f" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.614088 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.614563 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.623939 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-mg24f"] Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.715182 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mf5d4\" (UniqueName: \"kubernetes.io/projected/cf549f46-7597-4399-addc-3bf3f5b63f99-kube-api-access-mf5d4\") pod \"nova-cell1-cell-mapping-mg24f\" (UID: \"cf549f46-7597-4399-addc-3bf3f5b63f99\") " pod="openstack/nova-cell1-cell-mapping-mg24f" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.715253 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-scripts\") pod \"nova-cell1-cell-mapping-mg24f\" (UID: \"cf549f46-7597-4399-addc-3bf3f5b63f99\") " pod="openstack/nova-cell1-cell-mapping-mg24f" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.715277 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-config-data\") pod \"nova-cell1-cell-mapping-mg24f\" (UID: \"cf549f46-7597-4399-addc-3bf3f5b63f99\") " pod="openstack/nova-cell1-cell-mapping-mg24f" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.715358 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-mg24f\" (UID: \"cf549f46-7597-4399-addc-3bf3f5b63f99\") " pod="openstack/nova-cell1-cell-mapping-mg24f" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.816683 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-scripts\") pod \"nova-cell1-cell-mapping-mg24f\" (UID: \"cf549f46-7597-4399-addc-3bf3f5b63f99\") " pod="openstack/nova-cell1-cell-mapping-mg24f" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.816745 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-config-data\") pod \"nova-cell1-cell-mapping-mg24f\" (UID: \"cf549f46-7597-4399-addc-3bf3f5b63f99\") " pod="openstack/nova-cell1-cell-mapping-mg24f" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.816845 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-mg24f\" (UID: \"cf549f46-7597-4399-addc-3bf3f5b63f99\") " pod="openstack/nova-cell1-cell-mapping-mg24f" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.817037 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mf5d4\" (UniqueName: \"kubernetes.io/projected/cf549f46-7597-4399-addc-3bf3f5b63f99-kube-api-access-mf5d4\") pod \"nova-cell1-cell-mapping-mg24f\" (UID: \"cf549f46-7597-4399-addc-3bf3f5b63f99\") " pod="openstack/nova-cell1-cell-mapping-mg24f" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.823056 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-scripts\") pod \"nova-cell1-cell-mapping-mg24f\" (UID: \"cf549f46-7597-4399-addc-3bf3f5b63f99\") " pod="openstack/nova-cell1-cell-mapping-mg24f" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.830672 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-config-data\") pod \"nova-cell1-cell-mapping-mg24f\" (UID: \"cf549f46-7597-4399-addc-3bf3f5b63f99\") " pod="openstack/nova-cell1-cell-mapping-mg24f" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.831063 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-mg24f\" (UID: \"cf549f46-7597-4399-addc-3bf3f5b63f99\") " pod="openstack/nova-cell1-cell-mapping-mg24f" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.840624 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mf5d4\" (UniqueName: \"kubernetes.io/projected/cf549f46-7597-4399-addc-3bf3f5b63f99-kube-api-access-mf5d4\") pod \"nova-cell1-cell-mapping-mg24f\" (UID: \"cf549f46-7597-4399-addc-3bf3f5b63f99\") " pod="openstack/nova-cell1-cell-mapping-mg24f" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.950873 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mg24f" Feb 20 08:25:07 crc kubenswrapper[4948]: I0220 08:25:07.338229 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-mg24f"] Feb 20 08:25:07 crc kubenswrapper[4948]: I0220 08:25:07.403709 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mg24f" event={"ID":"cf549f46-7597-4399-addc-3bf3f5b63f99","Type":"ContainerStarted","Data":"f221fd286de80b0b0916709f7cbe440e6ef10e15899275765ef094ff94516225"} Feb 20 08:25:08 crc kubenswrapper[4948]: I0220 08:25:08.416705 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mg24f" event={"ID":"cf549f46-7597-4399-addc-3bf3f5b63f99","Type":"ContainerStarted","Data":"d68a0fb50da62e47bddf140b8e1628a7abf7e9ac68f7c4c4efaa6f9ccf62f179"} Feb 20 08:25:08 crc kubenswrapper[4948]: I0220 08:25:08.436714 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-mg24f" podStartSLOduration=2.436694909 podStartE2EDuration="2.436694909s" podCreationTimestamp="2026-02-20 08:25:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:25:08.434721291 +0000 UTC m=+1157.409216111" watchObservedRunningTime="2026-02-20 08:25:08.436694909 +0000 UTC m=+1157.411189729" Feb 20 08:25:08 crc kubenswrapper[4948]: I0220 08:25:08.718808 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 20 08:25:08 crc kubenswrapper[4948]: I0220 08:25:08.718849 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 20 08:25:09 crc kubenswrapper[4948]: I0220 08:25:09.739886 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="61657c77-bb29-4241-9d88-5e13bccd3f4e" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 20 08:25:09 crc kubenswrapper[4948]: I0220 08:25:09.740676 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="61657c77-bb29-4241-9d88-5e13bccd3f4e" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 20 08:25:12 crc kubenswrapper[4948]: I0220 08:25:12.452094 4948 generic.go:334] "Generic (PLEG): container finished" podID="cf549f46-7597-4399-addc-3bf3f5b63f99" containerID="d68a0fb50da62e47bddf140b8e1628a7abf7e9ac68f7c4c4efaa6f9ccf62f179" exitCode=0 Feb 20 08:25:12 crc kubenswrapper[4948]: I0220 08:25:12.452189 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mg24f" event={"ID":"cf549f46-7597-4399-addc-3bf3f5b63f99","Type":"ContainerDied","Data":"d68a0fb50da62e47bddf140b8e1628a7abf7e9ac68f7c4c4efaa6f9ccf62f179"} Feb 20 08:25:13 crc kubenswrapper[4948]: I0220 08:25:13.884118 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mg24f" Feb 20 08:25:13 crc kubenswrapper[4948]: I0220 08:25:13.974964 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mf5d4\" (UniqueName: \"kubernetes.io/projected/cf549f46-7597-4399-addc-3bf3f5b63f99-kube-api-access-mf5d4\") pod \"cf549f46-7597-4399-addc-3bf3f5b63f99\" (UID: \"cf549f46-7597-4399-addc-3bf3f5b63f99\") " Feb 20 08:25:13 crc kubenswrapper[4948]: I0220 08:25:13.975066 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-config-data\") pod \"cf549f46-7597-4399-addc-3bf3f5b63f99\" (UID: \"cf549f46-7597-4399-addc-3bf3f5b63f99\") " Feb 20 08:25:13 crc kubenswrapper[4948]: I0220 08:25:13.975117 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-combined-ca-bundle\") pod \"cf549f46-7597-4399-addc-3bf3f5b63f99\" (UID: \"cf549f46-7597-4399-addc-3bf3f5b63f99\") " Feb 20 08:25:13 crc kubenswrapper[4948]: I0220 08:25:13.975195 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-scripts\") pod \"cf549f46-7597-4399-addc-3bf3f5b63f99\" (UID: \"cf549f46-7597-4399-addc-3bf3f5b63f99\") " Feb 20 08:25:13 crc kubenswrapper[4948]: I0220 08:25:13.982448 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf549f46-7597-4399-addc-3bf3f5b63f99-kube-api-access-mf5d4" (OuterVolumeSpecName: "kube-api-access-mf5d4") pod "cf549f46-7597-4399-addc-3bf3f5b63f99" (UID: "cf549f46-7597-4399-addc-3bf3f5b63f99"). InnerVolumeSpecName "kube-api-access-mf5d4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:25:13 crc kubenswrapper[4948]: I0220 08:25:13.987695 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-scripts" (OuterVolumeSpecName: "scripts") pod "cf549f46-7597-4399-addc-3bf3f5b63f99" (UID: "cf549f46-7597-4399-addc-3bf3f5b63f99"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:25:14 crc kubenswrapper[4948]: I0220 08:25:14.002484 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-config-data" (OuterVolumeSpecName: "config-data") pod "cf549f46-7597-4399-addc-3bf3f5b63f99" (UID: "cf549f46-7597-4399-addc-3bf3f5b63f99"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:25:14 crc kubenswrapper[4948]: I0220 08:25:14.005161 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cf549f46-7597-4399-addc-3bf3f5b63f99" (UID: "cf549f46-7597-4399-addc-3bf3f5b63f99"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:25:14 crc kubenswrapper[4948]: I0220 08:25:14.077535 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mf5d4\" (UniqueName: \"kubernetes.io/projected/cf549f46-7597-4399-addc-3bf3f5b63f99-kube-api-access-mf5d4\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:14 crc kubenswrapper[4948]: I0220 08:25:14.077566 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:14 crc kubenswrapper[4948]: I0220 08:25:14.077576 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:14 crc kubenswrapper[4948]: I0220 08:25:14.077584 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:14 crc kubenswrapper[4948]: I0220 08:25:14.470859 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mg24f" event={"ID":"cf549f46-7597-4399-addc-3bf3f5b63f99","Type":"ContainerDied","Data":"f221fd286de80b0b0916709f7cbe440e6ef10e15899275765ef094ff94516225"} Feb 20 08:25:14 crc kubenswrapper[4948]: I0220 08:25:14.470893 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f221fd286de80b0b0916709f7cbe440e6ef10e15899275765ef094ff94516225" Feb 20 08:25:14 crc kubenswrapper[4948]: I0220 08:25:14.470960 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mg24f" Feb 20 08:25:14 crc kubenswrapper[4948]: I0220 08:25:14.675110 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 08:25:14 crc kubenswrapper[4948]: I0220 08:25:14.675735 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="a76350cb-da45-403e-8934-ef53b257ea27" containerName="nova-scheduler-scheduler" containerID="cri-o://7fdf44e32062d9617c830886caa9e40ee3055d0b238b4f4ec53507c904d02787" gracePeriod=30 Feb 20 08:25:14 crc kubenswrapper[4948]: I0220 08:25:14.695550 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 20 08:25:14 crc kubenswrapper[4948]: I0220 08:25:14.695829 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="61657c77-bb29-4241-9d88-5e13bccd3f4e" containerName="nova-api-log" containerID="cri-o://38deb2b3c4145711eaa3d46abb6b2827da68608a05ee0a36a26658e34f7dc7af" gracePeriod=30 Feb 20 08:25:14 crc kubenswrapper[4948]: I0220 08:25:14.695999 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="61657c77-bb29-4241-9d88-5e13bccd3f4e" containerName="nova-api-api" containerID="cri-o://185d2ffc7312a2c6e5587955ab1e83610f6970f92e1e567908cad389309483bc" gracePeriod=30 Feb 20 08:25:14 crc kubenswrapper[4948]: I0220 08:25:14.716701 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:25:14 crc kubenswrapper[4948]: I0220 08:25:14.717185 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="659ea49b-7e8f-4c95-ac69-ec743d0f95ed" containerName="nova-metadata-log" containerID="cri-o://63251b467317ee4b7966810a99bfea1f41b91c2ea1703852741d551304c8bdaf" gracePeriod=30 Feb 20 08:25:14 crc kubenswrapper[4948]: I0220 08:25:14.717302 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="659ea49b-7e8f-4c95-ac69-ec743d0f95ed" containerName="nova-metadata-metadata" containerID="cri-o://22196e9ec797dbe697aeeef9301c64792cbe1db93384c86a9e5389ca3fe652c9" gracePeriod=30 Feb 20 08:25:15 crc kubenswrapper[4948]: I0220 08:25:15.479841 4948 generic.go:334] "Generic (PLEG): container finished" podID="61657c77-bb29-4241-9d88-5e13bccd3f4e" containerID="38deb2b3c4145711eaa3d46abb6b2827da68608a05ee0a36a26658e34f7dc7af" exitCode=143 Feb 20 08:25:15 crc kubenswrapper[4948]: I0220 08:25:15.479917 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"61657c77-bb29-4241-9d88-5e13bccd3f4e","Type":"ContainerDied","Data":"38deb2b3c4145711eaa3d46abb6b2827da68608a05ee0a36a26658e34f7dc7af"} Feb 20 08:25:15 crc kubenswrapper[4948]: I0220 08:25:15.482020 4948 generic.go:334] "Generic (PLEG): container finished" podID="659ea49b-7e8f-4c95-ac69-ec743d0f95ed" containerID="63251b467317ee4b7966810a99bfea1f41b91c2ea1703852741d551304c8bdaf" exitCode=143 Feb 20 08:25:15 crc kubenswrapper[4948]: I0220 08:25:15.482062 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"659ea49b-7e8f-4c95-ac69-ec743d0f95ed","Type":"ContainerDied","Data":"63251b467317ee4b7966810a99bfea1f41b91c2ea1703852741d551304c8bdaf"} Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.097429 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.217952 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a76350cb-da45-403e-8934-ef53b257ea27-combined-ca-bundle\") pod \"a76350cb-da45-403e-8934-ef53b257ea27\" (UID: \"a76350cb-da45-403e-8934-ef53b257ea27\") " Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.218064 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a76350cb-da45-403e-8934-ef53b257ea27-config-data\") pod \"a76350cb-da45-403e-8934-ef53b257ea27\" (UID: \"a76350cb-da45-403e-8934-ef53b257ea27\") " Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.218179 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cb4nn\" (UniqueName: \"kubernetes.io/projected/a76350cb-da45-403e-8934-ef53b257ea27-kube-api-access-cb4nn\") pod \"a76350cb-da45-403e-8934-ef53b257ea27\" (UID: \"a76350cb-da45-403e-8934-ef53b257ea27\") " Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.236297 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a76350cb-da45-403e-8934-ef53b257ea27-kube-api-access-cb4nn" (OuterVolumeSpecName: "kube-api-access-cb4nn") pod "a76350cb-da45-403e-8934-ef53b257ea27" (UID: "a76350cb-da45-403e-8934-ef53b257ea27"). InnerVolumeSpecName "kube-api-access-cb4nn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.256980 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a76350cb-da45-403e-8934-ef53b257ea27-config-data" (OuterVolumeSpecName: "config-data") pod "a76350cb-da45-403e-8934-ef53b257ea27" (UID: "a76350cb-da45-403e-8934-ef53b257ea27"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.281489 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a76350cb-da45-403e-8934-ef53b257ea27-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a76350cb-da45-403e-8934-ef53b257ea27" (UID: "a76350cb-da45-403e-8934-ef53b257ea27"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.320138 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a76350cb-da45-403e-8934-ef53b257ea27-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.320299 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a76350cb-da45-403e-8934-ef53b257ea27-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.320358 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cb4nn\" (UniqueName: \"kubernetes.io/projected/a76350cb-da45-403e-8934-ef53b257ea27-kube-api-access-cb4nn\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.492265 4948 generic.go:334] "Generic (PLEG): container finished" podID="a76350cb-da45-403e-8934-ef53b257ea27" containerID="7fdf44e32062d9617c830886caa9e40ee3055d0b238b4f4ec53507c904d02787" exitCode=0 Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.492308 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.492327 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a76350cb-da45-403e-8934-ef53b257ea27","Type":"ContainerDied","Data":"7fdf44e32062d9617c830886caa9e40ee3055d0b238b4f4ec53507c904d02787"} Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.493537 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a76350cb-da45-403e-8934-ef53b257ea27","Type":"ContainerDied","Data":"6fbbcc32e0efcd622515f0c849bc2e7cae40139cc47cc3aeb9187b65bfe28ca5"} Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.493575 4948 scope.go:117] "RemoveContainer" containerID="7fdf44e32062d9617c830886caa9e40ee3055d0b238b4f4ec53507c904d02787" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.527126 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.530418 4948 scope.go:117] "RemoveContainer" containerID="7fdf44e32062d9617c830886caa9e40ee3055d0b238b4f4ec53507c904d02787" Feb 20 08:25:16 crc kubenswrapper[4948]: E0220 08:25:16.530831 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fdf44e32062d9617c830886caa9e40ee3055d0b238b4f4ec53507c904d02787\": container with ID starting with 7fdf44e32062d9617c830886caa9e40ee3055d0b238b4f4ec53507c904d02787 not found: ID does not exist" containerID="7fdf44e32062d9617c830886caa9e40ee3055d0b238b4f4ec53507c904d02787" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.530867 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fdf44e32062d9617c830886caa9e40ee3055d0b238b4f4ec53507c904d02787"} err="failed to get container status \"7fdf44e32062d9617c830886caa9e40ee3055d0b238b4f4ec53507c904d02787\": rpc error: code = NotFound desc = could not find container \"7fdf44e32062d9617c830886caa9e40ee3055d0b238b4f4ec53507c904d02787\": container with ID starting with 7fdf44e32062d9617c830886caa9e40ee3055d0b238b4f4ec53507c904d02787 not found: ID does not exist" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.537292 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.561359 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 08:25:16 crc kubenswrapper[4948]: E0220 08:25:16.561823 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a76350cb-da45-403e-8934-ef53b257ea27" containerName="nova-scheduler-scheduler" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.561842 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="a76350cb-da45-403e-8934-ef53b257ea27" containerName="nova-scheduler-scheduler" Feb 20 08:25:16 crc kubenswrapper[4948]: E0220 08:25:16.561867 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf549f46-7597-4399-addc-3bf3f5b63f99" containerName="nova-manage" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.561875 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf549f46-7597-4399-addc-3bf3f5b63f99" containerName="nova-manage" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.562174 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="a76350cb-da45-403e-8934-ef53b257ea27" containerName="nova-scheduler-scheduler" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.562226 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf549f46-7597-4399-addc-3bf3f5b63f99" containerName="nova-manage" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.563165 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.565324 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.582577 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.625241 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-js86x\" (UniqueName: \"kubernetes.io/projected/f114bc41-01a5-4955-97eb-7fcf139cc5a9-kube-api-access-js86x\") pod \"nova-scheduler-0\" (UID: \"f114bc41-01a5-4955-97eb-7fcf139cc5a9\") " pod="openstack/nova-scheduler-0" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.625280 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f114bc41-01a5-4955-97eb-7fcf139cc5a9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f114bc41-01a5-4955-97eb-7fcf139cc5a9\") " pod="openstack/nova-scheduler-0" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.625351 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f114bc41-01a5-4955-97eb-7fcf139cc5a9-config-data\") pod \"nova-scheduler-0\" (UID: \"f114bc41-01a5-4955-97eb-7fcf139cc5a9\") " pod="openstack/nova-scheduler-0" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.726551 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f114bc41-01a5-4955-97eb-7fcf139cc5a9-config-data\") pod \"nova-scheduler-0\" (UID: \"f114bc41-01a5-4955-97eb-7fcf139cc5a9\") " pod="openstack/nova-scheduler-0" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.726702 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-js86x\" (UniqueName: \"kubernetes.io/projected/f114bc41-01a5-4955-97eb-7fcf139cc5a9-kube-api-access-js86x\") pod \"nova-scheduler-0\" (UID: \"f114bc41-01a5-4955-97eb-7fcf139cc5a9\") " pod="openstack/nova-scheduler-0" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.726732 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f114bc41-01a5-4955-97eb-7fcf139cc5a9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f114bc41-01a5-4955-97eb-7fcf139cc5a9\") " pod="openstack/nova-scheduler-0" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.732713 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f114bc41-01a5-4955-97eb-7fcf139cc5a9-config-data\") pod \"nova-scheduler-0\" (UID: \"f114bc41-01a5-4955-97eb-7fcf139cc5a9\") " pod="openstack/nova-scheduler-0" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.732730 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f114bc41-01a5-4955-97eb-7fcf139cc5a9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f114bc41-01a5-4955-97eb-7fcf139cc5a9\") " pod="openstack/nova-scheduler-0" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.754832 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-js86x\" (UniqueName: \"kubernetes.io/projected/f114bc41-01a5-4955-97eb-7fcf139cc5a9-kube-api-access-js86x\") pod \"nova-scheduler-0\" (UID: \"f114bc41-01a5-4955-97eb-7fcf139cc5a9\") " pod="openstack/nova-scheduler-0" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.880688 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 20 08:25:17 crc kubenswrapper[4948]: I0220 08:25:17.324034 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 08:25:17 crc kubenswrapper[4948]: W0220 08:25:17.329498 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf114bc41_01a5_4955_97eb_7fcf139cc5a9.slice/crio-7221e125e11e0334cd672a971302d9b4b074d8307c59383d81cf365b5c4aac21 WatchSource:0}: Error finding container 7221e125e11e0334cd672a971302d9b4b074d8307c59383d81cf365b5c4aac21: Status 404 returned error can't find the container with id 7221e125e11e0334cd672a971302d9b4b074d8307c59383d81cf365b5c4aac21 Feb 20 08:25:17 crc kubenswrapper[4948]: I0220 08:25:17.507141 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f114bc41-01a5-4955-97eb-7fcf139cc5a9","Type":"ContainerStarted","Data":"7221e125e11e0334cd672a971302d9b4b074d8307c59383d81cf365b5c4aac21"} Feb 20 08:25:17 crc kubenswrapper[4948]: I0220 08:25:17.741700 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a76350cb-da45-403e-8934-ef53b257ea27" path="/var/lib/kubelet/pods/a76350cb-da45-403e-8934-ef53b257ea27/volumes" Feb 20 08:25:17 crc kubenswrapper[4948]: I0220 08:25:17.871244 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="659ea49b-7e8f-4c95-ac69-ec743d0f95ed" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.198:8775/\": read tcp 10.217.0.2:45064->10.217.0.198:8775: read: connection reset by peer" Feb 20 08:25:17 crc kubenswrapper[4948]: I0220 08:25:17.871257 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="659ea49b-7e8f-4c95-ac69-ec743d0f95ed" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.198:8775/\": read tcp 10.217.0.2:45078->10.217.0.198:8775: read: connection reset by peer" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.417927 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.423986 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.463320 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-config-data\") pod \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.463374 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-combined-ca-bundle\") pod \"61657c77-bb29-4241-9d88-5e13bccd3f4e\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.464242 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zr8fh\" (UniqueName: \"kubernetes.io/projected/61657c77-bb29-4241-9d88-5e13bccd3f4e-kube-api-access-zr8fh\") pod \"61657c77-bb29-4241-9d88-5e13bccd3f4e\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.464288 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-config-data\") pod \"61657c77-bb29-4241-9d88-5e13bccd3f4e\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.464355 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-logs\") pod \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.464398 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-internal-tls-certs\") pod \"61657c77-bb29-4241-9d88-5e13bccd3f4e\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.464440 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-combined-ca-bundle\") pod \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.464469 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-public-tls-certs\") pod \"61657c77-bb29-4241-9d88-5e13bccd3f4e\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.464523 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7ch5\" (UniqueName: \"kubernetes.io/projected/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-kube-api-access-p7ch5\") pod \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.464553 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61657c77-bb29-4241-9d88-5e13bccd3f4e-logs\") pod \"61657c77-bb29-4241-9d88-5e13bccd3f4e\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.464592 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-nova-metadata-tls-certs\") pod \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.470011 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61657c77-bb29-4241-9d88-5e13bccd3f4e-kube-api-access-zr8fh" (OuterVolumeSpecName: "kube-api-access-zr8fh") pod "61657c77-bb29-4241-9d88-5e13bccd3f4e" (UID: "61657c77-bb29-4241-9d88-5e13bccd3f4e"). InnerVolumeSpecName "kube-api-access-zr8fh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.470546 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61657c77-bb29-4241-9d88-5e13bccd3f4e-logs" (OuterVolumeSpecName: "logs") pod "61657c77-bb29-4241-9d88-5e13bccd3f4e" (UID: "61657c77-bb29-4241-9d88-5e13bccd3f4e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.472119 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-logs" (OuterVolumeSpecName: "logs") pod "659ea49b-7e8f-4c95-ac69-ec743d0f95ed" (UID: "659ea49b-7e8f-4c95-ac69-ec743d0f95ed"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.491826 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-kube-api-access-p7ch5" (OuterVolumeSpecName: "kube-api-access-p7ch5") pod "659ea49b-7e8f-4c95-ac69-ec743d0f95ed" (UID: "659ea49b-7e8f-4c95-ac69-ec743d0f95ed"). InnerVolumeSpecName "kube-api-access-p7ch5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.529769 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-config-data" (OuterVolumeSpecName: "config-data") pod "659ea49b-7e8f-4c95-ac69-ec743d0f95ed" (UID: "659ea49b-7e8f-4c95-ac69-ec743d0f95ed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.530054 4948 generic.go:334] "Generic (PLEG): container finished" podID="659ea49b-7e8f-4c95-ac69-ec743d0f95ed" containerID="22196e9ec797dbe697aeeef9301c64792cbe1db93384c86a9e5389ca3fe652c9" exitCode=0 Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.530120 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.530196 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"659ea49b-7e8f-4c95-ac69-ec743d0f95ed","Type":"ContainerDied","Data":"22196e9ec797dbe697aeeef9301c64792cbe1db93384c86a9e5389ca3fe652c9"} Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.530229 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"659ea49b-7e8f-4c95-ac69-ec743d0f95ed","Type":"ContainerDied","Data":"aaa4c4ee0564a915596564697d18943937054d1b42224076a88c2793a4afd7d0"} Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.530250 4948 scope.go:117] "RemoveContainer" containerID="22196e9ec797dbe697aeeef9301c64792cbe1db93384c86a9e5389ca3fe652c9" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.533752 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f114bc41-01a5-4955-97eb-7fcf139cc5a9","Type":"ContainerStarted","Data":"c760bba370b0c11b1caf2e537a3240838f2824510f92c38d77c5ed335aec7b26"} Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.542246 4948 generic.go:334] "Generic (PLEG): container finished" podID="61657c77-bb29-4241-9d88-5e13bccd3f4e" containerID="185d2ffc7312a2c6e5587955ab1e83610f6970f92e1e567908cad389309483bc" exitCode=0 Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.542316 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"61657c77-bb29-4241-9d88-5e13bccd3f4e","Type":"ContainerDied","Data":"185d2ffc7312a2c6e5587955ab1e83610f6970f92e1e567908cad389309483bc"} Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.542346 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"61657c77-bb29-4241-9d88-5e13bccd3f4e","Type":"ContainerDied","Data":"cff172040120f907a742a7013b9d3e9198e3f80ccb81cd1545dba19de50f6adb"} Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.542408 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.548196 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "61657c77-bb29-4241-9d88-5e13bccd3f4e" (UID: "61657c77-bb29-4241-9d88-5e13bccd3f4e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.558590 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.558568762 podStartE2EDuration="2.558568762s" podCreationTimestamp="2026-02-20 08:25:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:25:18.554158756 +0000 UTC m=+1167.528653596" watchObservedRunningTime="2026-02-20 08:25:18.558568762 +0000 UTC m=+1167.533063582" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.558888 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "659ea49b-7e8f-4c95-ac69-ec743d0f95ed" (UID: "659ea49b-7e8f-4c95-ac69-ec743d0f95ed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.566829 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.566860 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zr8fh\" (UniqueName: \"kubernetes.io/projected/61657c77-bb29-4241-9d88-5e13bccd3f4e-kube-api-access-zr8fh\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.566875 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.566887 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.566900 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7ch5\" (UniqueName: \"kubernetes.io/projected/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-kube-api-access-p7ch5\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.566911 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61657c77-bb29-4241-9d88-5e13bccd3f4e-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.566921 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.596542 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-config-data" (OuterVolumeSpecName: "config-data") pod "61657c77-bb29-4241-9d88-5e13bccd3f4e" (UID: "61657c77-bb29-4241-9d88-5e13bccd3f4e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.597665 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "61657c77-bb29-4241-9d88-5e13bccd3f4e" (UID: "61657c77-bb29-4241-9d88-5e13bccd3f4e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.599545 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "61657c77-bb29-4241-9d88-5e13bccd3f4e" (UID: "61657c77-bb29-4241-9d88-5e13bccd3f4e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.623883 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "659ea49b-7e8f-4c95-ac69-ec743d0f95ed" (UID: "659ea49b-7e8f-4c95-ac69-ec743d0f95ed"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.626180 4948 scope.go:117] "RemoveContainer" containerID="63251b467317ee4b7966810a99bfea1f41b91c2ea1703852741d551304c8bdaf" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.645144 4948 scope.go:117] "RemoveContainer" containerID="22196e9ec797dbe697aeeef9301c64792cbe1db93384c86a9e5389ca3fe652c9" Feb 20 08:25:18 crc kubenswrapper[4948]: E0220 08:25:18.646478 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22196e9ec797dbe697aeeef9301c64792cbe1db93384c86a9e5389ca3fe652c9\": container with ID starting with 22196e9ec797dbe697aeeef9301c64792cbe1db93384c86a9e5389ca3fe652c9 not found: ID does not exist" containerID="22196e9ec797dbe697aeeef9301c64792cbe1db93384c86a9e5389ca3fe652c9" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.646586 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22196e9ec797dbe697aeeef9301c64792cbe1db93384c86a9e5389ca3fe652c9"} err="failed to get container status \"22196e9ec797dbe697aeeef9301c64792cbe1db93384c86a9e5389ca3fe652c9\": rpc error: code = NotFound desc = could not find container \"22196e9ec797dbe697aeeef9301c64792cbe1db93384c86a9e5389ca3fe652c9\": container with ID starting with 22196e9ec797dbe697aeeef9301c64792cbe1db93384c86a9e5389ca3fe652c9 not found: ID does not exist" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.646718 4948 scope.go:117] "RemoveContainer" containerID="63251b467317ee4b7966810a99bfea1f41b91c2ea1703852741d551304c8bdaf" Feb 20 08:25:18 crc kubenswrapper[4948]: E0220 08:25:18.647183 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63251b467317ee4b7966810a99bfea1f41b91c2ea1703852741d551304c8bdaf\": container with ID starting with 63251b467317ee4b7966810a99bfea1f41b91c2ea1703852741d551304c8bdaf not found: ID does not exist" containerID="63251b467317ee4b7966810a99bfea1f41b91c2ea1703852741d551304c8bdaf" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.647258 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63251b467317ee4b7966810a99bfea1f41b91c2ea1703852741d551304c8bdaf"} err="failed to get container status \"63251b467317ee4b7966810a99bfea1f41b91c2ea1703852741d551304c8bdaf\": rpc error: code = NotFound desc = could not find container \"63251b467317ee4b7966810a99bfea1f41b91c2ea1703852741d551304c8bdaf\": container with ID starting with 63251b467317ee4b7966810a99bfea1f41b91c2ea1703852741d551304c8bdaf not found: ID does not exist" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.647371 4948 scope.go:117] "RemoveContainer" containerID="185d2ffc7312a2c6e5587955ab1e83610f6970f92e1e567908cad389309483bc" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.663963 4948 scope.go:117] "RemoveContainer" containerID="38deb2b3c4145711eaa3d46abb6b2827da68608a05ee0a36a26658e34f7dc7af" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.668851 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.668879 4948 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.668889 4948 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.668899 4948 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.685050 4948 scope.go:117] "RemoveContainer" containerID="185d2ffc7312a2c6e5587955ab1e83610f6970f92e1e567908cad389309483bc" Feb 20 08:25:18 crc kubenswrapper[4948]: E0220 08:25:18.685876 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"185d2ffc7312a2c6e5587955ab1e83610f6970f92e1e567908cad389309483bc\": container with ID starting with 185d2ffc7312a2c6e5587955ab1e83610f6970f92e1e567908cad389309483bc not found: ID does not exist" containerID="185d2ffc7312a2c6e5587955ab1e83610f6970f92e1e567908cad389309483bc" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.685942 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"185d2ffc7312a2c6e5587955ab1e83610f6970f92e1e567908cad389309483bc"} err="failed to get container status \"185d2ffc7312a2c6e5587955ab1e83610f6970f92e1e567908cad389309483bc\": rpc error: code = NotFound desc = could not find container \"185d2ffc7312a2c6e5587955ab1e83610f6970f92e1e567908cad389309483bc\": container with ID starting with 185d2ffc7312a2c6e5587955ab1e83610f6970f92e1e567908cad389309483bc not found: ID does not exist" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.686006 4948 scope.go:117] "RemoveContainer" containerID="38deb2b3c4145711eaa3d46abb6b2827da68608a05ee0a36a26658e34f7dc7af" Feb 20 08:25:18 crc kubenswrapper[4948]: E0220 08:25:18.686358 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38deb2b3c4145711eaa3d46abb6b2827da68608a05ee0a36a26658e34f7dc7af\": container with ID starting with 38deb2b3c4145711eaa3d46abb6b2827da68608a05ee0a36a26658e34f7dc7af not found: ID does not exist" containerID="38deb2b3c4145711eaa3d46abb6b2827da68608a05ee0a36a26658e34f7dc7af" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.686401 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38deb2b3c4145711eaa3d46abb6b2827da68608a05ee0a36a26658e34f7dc7af"} err="failed to get container status \"38deb2b3c4145711eaa3d46abb6b2827da68608a05ee0a36a26658e34f7dc7af\": rpc error: code = NotFound desc = could not find container \"38deb2b3c4145711eaa3d46abb6b2827da68608a05ee0a36a26658e34f7dc7af\": container with ID starting with 38deb2b3c4145711eaa3d46abb6b2827da68608a05ee0a36a26658e34f7dc7af not found: ID does not exist" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.880234 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.894444 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.905916 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:25:18 crc kubenswrapper[4948]: E0220 08:25:18.906420 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="659ea49b-7e8f-4c95-ac69-ec743d0f95ed" containerName="nova-metadata-metadata" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.906447 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="659ea49b-7e8f-4c95-ac69-ec743d0f95ed" containerName="nova-metadata-metadata" Feb 20 08:25:18 crc kubenswrapper[4948]: E0220 08:25:18.906470 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61657c77-bb29-4241-9d88-5e13bccd3f4e" containerName="nova-api-log" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.906480 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="61657c77-bb29-4241-9d88-5e13bccd3f4e" containerName="nova-api-log" Feb 20 08:25:18 crc kubenswrapper[4948]: E0220 08:25:18.906493 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61657c77-bb29-4241-9d88-5e13bccd3f4e" containerName="nova-api-api" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.906501 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="61657c77-bb29-4241-9d88-5e13bccd3f4e" containerName="nova-api-api" Feb 20 08:25:18 crc kubenswrapper[4948]: E0220 08:25:18.906531 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="659ea49b-7e8f-4c95-ac69-ec743d0f95ed" containerName="nova-metadata-log" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.906540 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="659ea49b-7e8f-4c95-ac69-ec743d0f95ed" containerName="nova-metadata-log" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.906733 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="659ea49b-7e8f-4c95-ac69-ec743d0f95ed" containerName="nova-metadata-log" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.906761 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="659ea49b-7e8f-4c95-ac69-ec743d0f95ed" containerName="nova-metadata-metadata" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.906775 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="61657c77-bb29-4241-9d88-5e13bccd3f4e" containerName="nova-api-log" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.906791 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="61657c77-bb29-4241-9d88-5e13bccd3f4e" containerName="nova-api-api" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.908046 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.911192 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.912551 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.915696 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.924333 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.936099 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.944383 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.946426 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.949254 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.949720 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.960968 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.972277 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.974487 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-public-tls-certs\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.974535 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.974585 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-logs\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.974640 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.974679 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59965363-9971-4130-bb5a-5fcf5a44e2d5-logs\") pod \"nova-metadata-0\" (UID: \"59965363-9971-4130-bb5a-5fcf5a44e2d5\") " pod="openstack/nova-metadata-0" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.974699 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/59965363-9971-4130-bb5a-5fcf5a44e2d5-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"59965363-9971-4130-bb5a-5fcf5a44e2d5\") " pod="openstack/nova-metadata-0" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.974738 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhr8w\" (UniqueName: \"kubernetes.io/projected/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-kube-api-access-zhr8w\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.974806 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59965363-9971-4130-bb5a-5fcf5a44e2d5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"59965363-9971-4130-bb5a-5fcf5a44e2d5\") " pod="openstack/nova-metadata-0" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.974848 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59965363-9971-4130-bb5a-5fcf5a44e2d5-config-data\") pod \"nova-metadata-0\" (UID: \"59965363-9971-4130-bb5a-5fcf5a44e2d5\") " pod="openstack/nova-metadata-0" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.974885 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-config-data\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.974916 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sqtf\" (UniqueName: \"kubernetes.io/projected/59965363-9971-4130-bb5a-5fcf5a44e2d5-kube-api-access-5sqtf\") pod \"nova-metadata-0\" (UID: \"59965363-9971-4130-bb5a-5fcf5a44e2d5\") " pod="openstack/nova-metadata-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.076982 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-public-tls-certs\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.077041 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.077093 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-logs\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.077130 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.077167 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59965363-9971-4130-bb5a-5fcf5a44e2d5-logs\") pod \"nova-metadata-0\" (UID: \"59965363-9971-4130-bb5a-5fcf5a44e2d5\") " pod="openstack/nova-metadata-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.077189 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/59965363-9971-4130-bb5a-5fcf5a44e2d5-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"59965363-9971-4130-bb5a-5fcf5a44e2d5\") " pod="openstack/nova-metadata-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.077229 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhr8w\" (UniqueName: \"kubernetes.io/projected/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-kube-api-access-zhr8w\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.077293 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59965363-9971-4130-bb5a-5fcf5a44e2d5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"59965363-9971-4130-bb5a-5fcf5a44e2d5\") " pod="openstack/nova-metadata-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.077327 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59965363-9971-4130-bb5a-5fcf5a44e2d5-config-data\") pod \"nova-metadata-0\" (UID: \"59965363-9971-4130-bb5a-5fcf5a44e2d5\") " pod="openstack/nova-metadata-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.077363 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-config-data\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.077396 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sqtf\" (UniqueName: \"kubernetes.io/projected/59965363-9971-4130-bb5a-5fcf5a44e2d5-kube-api-access-5sqtf\") pod \"nova-metadata-0\" (UID: \"59965363-9971-4130-bb5a-5fcf5a44e2d5\") " pod="openstack/nova-metadata-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.077543 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-logs\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.078098 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59965363-9971-4130-bb5a-5fcf5a44e2d5-logs\") pod \"nova-metadata-0\" (UID: \"59965363-9971-4130-bb5a-5fcf5a44e2d5\") " pod="openstack/nova-metadata-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.080016 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-public-tls-certs\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.080837 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-config-data\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.080909 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59965363-9971-4130-bb5a-5fcf5a44e2d5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"59965363-9971-4130-bb5a-5fcf5a44e2d5\") " pod="openstack/nova-metadata-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.081584 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59965363-9971-4130-bb5a-5fcf5a44e2d5-config-data\") pod \"nova-metadata-0\" (UID: \"59965363-9971-4130-bb5a-5fcf5a44e2d5\") " pod="openstack/nova-metadata-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.081910 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.083376 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.084056 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/59965363-9971-4130-bb5a-5fcf5a44e2d5-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"59965363-9971-4130-bb5a-5fcf5a44e2d5\") " pod="openstack/nova-metadata-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.093019 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhr8w\" (UniqueName: \"kubernetes.io/projected/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-kube-api-access-zhr8w\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.094819 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sqtf\" (UniqueName: \"kubernetes.io/projected/59965363-9971-4130-bb5a-5fcf5a44e2d5-kube-api-access-5sqtf\") pod \"nova-metadata-0\" (UID: \"59965363-9971-4130-bb5a-5fcf5a44e2d5\") " pod="openstack/nova-metadata-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.226041 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.270260 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 08:25:19 crc kubenswrapper[4948]: W0220 08:25:19.717811 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod59965363_9971_4130_bb5a_5fcf5a44e2d5.slice/crio-76ceb8ccdbd76e943e8b1df093a8e84d25b146d383c80a19a3539cb59aa2e930 WatchSource:0}: Error finding container 76ceb8ccdbd76e943e8b1df093a8e84d25b146d383c80a19a3539cb59aa2e930: Status 404 returned error can't find the container with id 76ceb8ccdbd76e943e8b1df093a8e84d25b146d383c80a19a3539cb59aa2e930 Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.719186 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.762291 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61657c77-bb29-4241-9d88-5e13bccd3f4e" path="/var/lib/kubelet/pods/61657c77-bb29-4241-9d88-5e13bccd3f4e/volumes" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.762964 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="659ea49b-7e8f-4c95-ac69-ec743d0f95ed" path="/var/lib/kubelet/pods/659ea49b-7e8f-4c95-ac69-ec743d0f95ed/volumes" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.789995 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 20 08:25:19 crc kubenswrapper[4948]: W0220 08:25:19.801905 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod354f9b2f_9eaa_4e47_b214_0fd895f90c4c.slice/crio-5b932a3154a482569582f2a63f0553401470a58c072e8c1eab52792d5c807f8f WatchSource:0}: Error finding container 5b932a3154a482569582f2a63f0553401470a58c072e8c1eab52792d5c807f8f: Status 404 returned error can't find the container with id 5b932a3154a482569582f2a63f0553401470a58c072e8c1eab52792d5c807f8f Feb 20 08:25:20 crc kubenswrapper[4948]: I0220 08:25:20.569030 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"59965363-9971-4130-bb5a-5fcf5a44e2d5","Type":"ContainerStarted","Data":"6671f87799658c87c9825b2bf7885aedd1ec50455ebb820ad3145ff0cd4079bb"} Feb 20 08:25:20 crc kubenswrapper[4948]: I0220 08:25:20.569284 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"59965363-9971-4130-bb5a-5fcf5a44e2d5","Type":"ContainerStarted","Data":"c27f81fe70c91cfb56c95520ee8e8ab8efccada3d580de7a3beb5a071d8d9267"} Feb 20 08:25:20 crc kubenswrapper[4948]: I0220 08:25:20.569293 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"59965363-9971-4130-bb5a-5fcf5a44e2d5","Type":"ContainerStarted","Data":"76ceb8ccdbd76e943e8b1df093a8e84d25b146d383c80a19a3539cb59aa2e930"} Feb 20 08:25:20 crc kubenswrapper[4948]: I0220 08:25:20.572086 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"354f9b2f-9eaa-4e47-b214-0fd895f90c4c","Type":"ContainerStarted","Data":"fc3c49087b8c230b070d30b6ec36f8bd733f16769615b95321e2bb087eea68f1"} Feb 20 08:25:20 crc kubenswrapper[4948]: I0220 08:25:20.572115 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"354f9b2f-9eaa-4e47-b214-0fd895f90c4c","Type":"ContainerStarted","Data":"671fd39418c48112cbf952091c772751dc7a1df73ad2c05de77d69960fd1c5b5"} Feb 20 08:25:20 crc kubenswrapper[4948]: I0220 08:25:20.572124 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"354f9b2f-9eaa-4e47-b214-0fd895f90c4c","Type":"ContainerStarted","Data":"5b932a3154a482569582f2a63f0553401470a58c072e8c1eab52792d5c807f8f"} Feb 20 08:25:20 crc kubenswrapper[4948]: I0220 08:25:20.609573 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.609557508 podStartE2EDuration="2.609557508s" podCreationTimestamp="2026-02-20 08:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:25:20.604117086 +0000 UTC m=+1169.578611916" watchObservedRunningTime="2026-02-20 08:25:20.609557508 +0000 UTC m=+1169.584052318" Feb 20 08:25:20 crc kubenswrapper[4948]: I0220 08:25:20.639360 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.639344946 podStartE2EDuration="2.639344946s" podCreationTimestamp="2026-02-20 08:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:25:20.633591408 +0000 UTC m=+1169.608086228" watchObservedRunningTime="2026-02-20 08:25:20.639344946 +0000 UTC m=+1169.613839766" Feb 20 08:25:21 crc kubenswrapper[4948]: I0220 08:25:21.881303 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 20 08:25:24 crc kubenswrapper[4948]: I0220 08:25:24.226791 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 20 08:25:24 crc kubenswrapper[4948]: I0220 08:25:24.227378 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 20 08:25:26 crc kubenswrapper[4948]: I0220 08:25:26.881581 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 20 08:25:26 crc kubenswrapper[4948]: I0220 08:25:26.924538 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 20 08:25:27 crc kubenswrapper[4948]: I0220 08:25:27.693124 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 20 08:25:28 crc kubenswrapper[4948]: I0220 08:25:28.753356 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 20 08:25:29 crc kubenswrapper[4948]: I0220 08:25:29.226338 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 20 08:25:29 crc kubenswrapper[4948]: I0220 08:25:29.226676 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 20 08:25:29 crc kubenswrapper[4948]: I0220 08:25:29.271431 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 20 08:25:29 crc kubenswrapper[4948]: I0220 08:25:29.271517 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 20 08:25:30 crc kubenswrapper[4948]: I0220 08:25:30.241174 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="59965363-9971-4130-bb5a-5fcf5a44e2d5" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.206:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 20 08:25:30 crc kubenswrapper[4948]: I0220 08:25:30.241203 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="59965363-9971-4130-bb5a-5fcf5a44e2d5" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.206:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 20 08:25:30 crc kubenswrapper[4948]: I0220 08:25:30.287228 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="354f9b2f-9eaa-4e47-b214-0fd895f90c4c" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.207:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 20 08:25:30 crc kubenswrapper[4948]: I0220 08:25:30.287277 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="354f9b2f-9eaa-4e47-b214-0fd895f90c4c" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.207:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 20 08:25:32 crc kubenswrapper[4948]: I0220 08:25:32.854739 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 20 08:25:32 crc kubenswrapper[4948]: I0220 08:25:32.856467 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="0dfbb200-e444-4895-a410-e51cbb1fc112" containerName="kube-state-metrics" containerID="cri-o://20c91950ce09f61ade4d935382ea88014ac2e6914fdd0f747d53e6577cf780a0" gracePeriod=30 Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.348826 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.460746 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqgmb\" (UniqueName: \"kubernetes.io/projected/0dfbb200-e444-4895-a410-e51cbb1fc112-kube-api-access-rqgmb\") pod \"0dfbb200-e444-4895-a410-e51cbb1fc112\" (UID: \"0dfbb200-e444-4895-a410-e51cbb1fc112\") " Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.475265 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dfbb200-e444-4895-a410-e51cbb1fc112-kube-api-access-rqgmb" (OuterVolumeSpecName: "kube-api-access-rqgmb") pod "0dfbb200-e444-4895-a410-e51cbb1fc112" (UID: "0dfbb200-e444-4895-a410-e51cbb1fc112"). InnerVolumeSpecName "kube-api-access-rqgmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.563006 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqgmb\" (UniqueName: \"kubernetes.io/projected/0dfbb200-e444-4895-a410-e51cbb1fc112-kube-api-access-rqgmb\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.698311 4948 generic.go:334] "Generic (PLEG): container finished" podID="0dfbb200-e444-4895-a410-e51cbb1fc112" containerID="20c91950ce09f61ade4d935382ea88014ac2e6914fdd0f747d53e6577cf780a0" exitCode=2 Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.698354 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"0dfbb200-e444-4895-a410-e51cbb1fc112","Type":"ContainerDied","Data":"20c91950ce09f61ade4d935382ea88014ac2e6914fdd0f747d53e6577cf780a0"} Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.698379 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"0dfbb200-e444-4895-a410-e51cbb1fc112","Type":"ContainerDied","Data":"8d55ac91670695a37a460e364e6527e261fc751cd8cb9dd56d60bf1aa04e7514"} Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.698381 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.698406 4948 scope.go:117] "RemoveContainer" containerID="20c91950ce09f61ade4d935382ea88014ac2e6914fdd0f747d53e6577cf780a0" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.729231 4948 scope.go:117] "RemoveContainer" containerID="20c91950ce09f61ade4d935382ea88014ac2e6914fdd0f747d53e6577cf780a0" Feb 20 08:25:33 crc kubenswrapper[4948]: E0220 08:25:33.729845 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20c91950ce09f61ade4d935382ea88014ac2e6914fdd0f747d53e6577cf780a0\": container with ID starting with 20c91950ce09f61ade4d935382ea88014ac2e6914fdd0f747d53e6577cf780a0 not found: ID does not exist" containerID="20c91950ce09f61ade4d935382ea88014ac2e6914fdd0f747d53e6577cf780a0" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.729886 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20c91950ce09f61ade4d935382ea88014ac2e6914fdd0f747d53e6577cf780a0"} err="failed to get container status \"20c91950ce09f61ade4d935382ea88014ac2e6914fdd0f747d53e6577cf780a0\": rpc error: code = NotFound desc = could not find container \"20c91950ce09f61ade4d935382ea88014ac2e6914fdd0f747d53e6577cf780a0\": container with ID starting with 20c91950ce09f61ade4d935382ea88014ac2e6914fdd0f747d53e6577cf780a0 not found: ID does not exist" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.747625 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.769020 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.781656 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 20 08:25:33 crc kubenswrapper[4948]: E0220 08:25:33.782241 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dfbb200-e444-4895-a410-e51cbb1fc112" containerName="kube-state-metrics" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.782269 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dfbb200-e444-4895-a410-e51cbb1fc112" containerName="kube-state-metrics" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.782555 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dfbb200-e444-4895-a410-e51cbb1fc112" containerName="kube-state-metrics" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.783407 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.785153 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.788687 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.793581 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.868693 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdm4g\" (UniqueName: \"kubernetes.io/projected/24bf0fb4-81db-4955-8b5f-92c0f6b0856b-kube-api-access-sdm4g\") pod \"kube-state-metrics-0\" (UID: \"24bf0fb4-81db-4955-8b5f-92c0f6b0856b\") " pod="openstack/kube-state-metrics-0" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.868738 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/24bf0fb4-81db-4955-8b5f-92c0f6b0856b-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"24bf0fb4-81db-4955-8b5f-92c0f6b0856b\") " pod="openstack/kube-state-metrics-0" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.868951 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24bf0fb4-81db-4955-8b5f-92c0f6b0856b-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"24bf0fb4-81db-4955-8b5f-92c0f6b0856b\") " pod="openstack/kube-state-metrics-0" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.869265 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/24bf0fb4-81db-4955-8b5f-92c0f6b0856b-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"24bf0fb4-81db-4955-8b5f-92c0f6b0856b\") " pod="openstack/kube-state-metrics-0" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.970792 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24bf0fb4-81db-4955-8b5f-92c0f6b0856b-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"24bf0fb4-81db-4955-8b5f-92c0f6b0856b\") " pod="openstack/kube-state-metrics-0" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.970858 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/24bf0fb4-81db-4955-8b5f-92c0f6b0856b-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"24bf0fb4-81db-4955-8b5f-92c0f6b0856b\") " pod="openstack/kube-state-metrics-0" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.970935 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdm4g\" (UniqueName: \"kubernetes.io/projected/24bf0fb4-81db-4955-8b5f-92c0f6b0856b-kube-api-access-sdm4g\") pod \"kube-state-metrics-0\" (UID: \"24bf0fb4-81db-4955-8b5f-92c0f6b0856b\") " pod="openstack/kube-state-metrics-0" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.971334 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/24bf0fb4-81db-4955-8b5f-92c0f6b0856b-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"24bf0fb4-81db-4955-8b5f-92c0f6b0856b\") " pod="openstack/kube-state-metrics-0" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.974689 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/24bf0fb4-81db-4955-8b5f-92c0f6b0856b-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"24bf0fb4-81db-4955-8b5f-92c0f6b0856b\") " pod="openstack/kube-state-metrics-0" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.974823 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/24bf0fb4-81db-4955-8b5f-92c0f6b0856b-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"24bf0fb4-81db-4955-8b5f-92c0f6b0856b\") " pod="openstack/kube-state-metrics-0" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.975323 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24bf0fb4-81db-4955-8b5f-92c0f6b0856b-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"24bf0fb4-81db-4955-8b5f-92c0f6b0856b\") " pod="openstack/kube-state-metrics-0" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.987772 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdm4g\" (UniqueName: \"kubernetes.io/projected/24bf0fb4-81db-4955-8b5f-92c0f6b0856b-kube-api-access-sdm4g\") pod \"kube-state-metrics-0\" (UID: \"24bf0fb4-81db-4955-8b5f-92c0f6b0856b\") " pod="openstack/kube-state-metrics-0" Feb 20 08:25:34 crc kubenswrapper[4948]: I0220 08:25:34.107278 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 20 08:25:34 crc kubenswrapper[4948]: I0220 08:25:34.626481 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 20 08:25:34 crc kubenswrapper[4948]: I0220 08:25:34.707687 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"24bf0fb4-81db-4955-8b5f-92c0f6b0856b","Type":"ContainerStarted","Data":"e2d46f0550c278b036f36f0b63c9d548764224188f6468a3d293b475c6b6027b"} Feb 20 08:25:34 crc kubenswrapper[4948]: I0220 08:25:34.788221 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:25:34 crc kubenswrapper[4948]: I0220 08:25:34.788658 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerName="proxy-httpd" containerID="cri-o://a1bbb73bd50190b5b08c60773344adf5c06efb89abf70ae7646fd73534b398b9" gracePeriod=30 Feb 20 08:25:34 crc kubenswrapper[4948]: I0220 08:25:34.788823 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerName="sg-core" containerID="cri-o://e04f1d97ea889929a68ccc1a2406ae3771d7b5ea9af14caaec52d943991404e0" gracePeriod=30 Feb 20 08:25:34 crc kubenswrapper[4948]: I0220 08:25:34.788624 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerName="ceilometer-central-agent" containerID="cri-o://d03b827c5401b1d3755870ce6ce51a7201888c77eda8df6a4497ae437d66e88a" gracePeriod=30 Feb 20 08:25:34 crc kubenswrapper[4948]: I0220 08:25:34.789011 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerName="ceilometer-notification-agent" containerID="cri-o://43e2a82447778e081804b3e50b6f485e50305ec4eb1bbe2f231f259d11b87542" gracePeriod=30 Feb 20 08:25:35 crc kubenswrapper[4948]: I0220 08:25:35.736790 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"24bf0fb4-81db-4955-8b5f-92c0f6b0856b","Type":"ContainerStarted","Data":"8bbce6c2092a12dfd665eb6fc8e2c1e2aec7b889df488872597a31f777fdc8d3"} Feb 20 08:25:35 crc kubenswrapper[4948]: I0220 08:25:35.748244 4948 generic.go:334] "Generic (PLEG): container finished" podID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerID="a1bbb73bd50190b5b08c60773344adf5c06efb89abf70ae7646fd73534b398b9" exitCode=0 Feb 20 08:25:35 crc kubenswrapper[4948]: I0220 08:25:35.748274 4948 generic.go:334] "Generic (PLEG): container finished" podID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerID="e04f1d97ea889929a68ccc1a2406ae3771d7b5ea9af14caaec52d943991404e0" exitCode=2 Feb 20 08:25:35 crc kubenswrapper[4948]: I0220 08:25:35.748281 4948 generic.go:334] "Generic (PLEG): container finished" podID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerID="d03b827c5401b1d3755870ce6ce51a7201888c77eda8df6a4497ae437d66e88a" exitCode=0 Feb 20 08:25:35 crc kubenswrapper[4948]: I0220 08:25:35.761951 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0dfbb200-e444-4895-a410-e51cbb1fc112" path="/var/lib/kubelet/pods/0dfbb200-e444-4895-a410-e51cbb1fc112/volumes" Feb 20 08:25:35 crc kubenswrapper[4948]: I0220 08:25:35.765435 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 20 08:25:35 crc kubenswrapper[4948]: I0220 08:25:35.765604 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8","Type":"ContainerDied","Data":"a1bbb73bd50190b5b08c60773344adf5c06efb89abf70ae7646fd73534b398b9"} Feb 20 08:25:35 crc kubenswrapper[4948]: I0220 08:25:35.765739 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8","Type":"ContainerDied","Data":"e04f1d97ea889929a68ccc1a2406ae3771d7b5ea9af14caaec52d943991404e0"} Feb 20 08:25:35 crc kubenswrapper[4948]: I0220 08:25:35.765855 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8","Type":"ContainerDied","Data":"d03b827c5401b1d3755870ce6ce51a7201888c77eda8df6a4497ae437d66e88a"} Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.709094 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.738134 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.3084262349999998 podStartE2EDuration="3.738116545s" podCreationTimestamp="2026-02-20 08:25:33 +0000 UTC" firstStartedPulling="2026-02-20 08:25:34.611939256 +0000 UTC m=+1183.586434076" lastFinishedPulling="2026-02-20 08:25:35.041629566 +0000 UTC m=+1184.016124386" observedRunningTime="2026-02-20 08:25:35.761728813 +0000 UTC m=+1184.736223633" watchObservedRunningTime="2026-02-20 08:25:36.738116545 +0000 UTC m=+1185.712611365" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.762218 4948 generic.go:334] "Generic (PLEG): container finished" podID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerID="43e2a82447778e081804b3e50b6f485e50305ec4eb1bbe2f231f259d11b87542" exitCode=0 Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.762281 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8","Type":"ContainerDied","Data":"43e2a82447778e081804b3e50b6f485e50305ec4eb1bbe2f231f259d11b87542"} Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.762336 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8","Type":"ContainerDied","Data":"48f0f09932c5f951969d8a7a57693456f1c18049262597223bdbf2f58c0804b2"} Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.762330 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.762365 4948 scope.go:117] "RemoveContainer" containerID="a1bbb73bd50190b5b08c60773344adf5c06efb89abf70ae7646fd73534b398b9" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.787782 4948 scope.go:117] "RemoveContainer" containerID="e04f1d97ea889929a68ccc1a2406ae3771d7b5ea9af14caaec52d943991404e0" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.805247 4948 scope.go:117] "RemoveContainer" containerID="43e2a82447778e081804b3e50b6f485e50305ec4eb1bbe2f231f259d11b87542" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.824650 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-combined-ca-bundle\") pod \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.824809 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-run-httpd\") pod \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.824949 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-config-data\") pod \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.825048 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-sg-core-conf-yaml\") pod \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.825107 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-scripts\") pod \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.825129 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwwjr\" (UniqueName: \"kubernetes.io/projected/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-kube-api-access-mwwjr\") pod \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.825150 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-log-httpd\") pod \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.825222 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" (UID: "be0e09bc-edd8-4ea9-8a59-ab630ed91cd8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.825567 4948 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.827518 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" (UID: "be0e09bc-edd8-4ea9-8a59-ab630ed91cd8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.833331 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-scripts" (OuterVolumeSpecName: "scripts") pod "be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" (UID: "be0e09bc-edd8-4ea9-8a59-ab630ed91cd8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.833573 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-kube-api-access-mwwjr" (OuterVolumeSpecName: "kube-api-access-mwwjr") pod "be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" (UID: "be0e09bc-edd8-4ea9-8a59-ab630ed91cd8"). InnerVolumeSpecName "kube-api-access-mwwjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.848926 4948 scope.go:117] "RemoveContainer" containerID="d03b827c5401b1d3755870ce6ce51a7201888c77eda8df6a4497ae437d66e88a" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.853109 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" (UID: "be0e09bc-edd8-4ea9-8a59-ab630ed91cd8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.895847 4948 scope.go:117] "RemoveContainer" containerID="a1bbb73bd50190b5b08c60773344adf5c06efb89abf70ae7646fd73534b398b9" Feb 20 08:25:36 crc kubenswrapper[4948]: E0220 08:25:36.900517 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1bbb73bd50190b5b08c60773344adf5c06efb89abf70ae7646fd73534b398b9\": container with ID starting with a1bbb73bd50190b5b08c60773344adf5c06efb89abf70ae7646fd73534b398b9 not found: ID does not exist" containerID="a1bbb73bd50190b5b08c60773344adf5c06efb89abf70ae7646fd73534b398b9" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.900548 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1bbb73bd50190b5b08c60773344adf5c06efb89abf70ae7646fd73534b398b9"} err="failed to get container status \"a1bbb73bd50190b5b08c60773344adf5c06efb89abf70ae7646fd73534b398b9\": rpc error: code = NotFound desc = could not find container \"a1bbb73bd50190b5b08c60773344adf5c06efb89abf70ae7646fd73534b398b9\": container with ID starting with a1bbb73bd50190b5b08c60773344adf5c06efb89abf70ae7646fd73534b398b9 not found: ID does not exist" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.900568 4948 scope.go:117] "RemoveContainer" containerID="e04f1d97ea889929a68ccc1a2406ae3771d7b5ea9af14caaec52d943991404e0" Feb 20 08:25:36 crc kubenswrapper[4948]: E0220 08:25:36.902523 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e04f1d97ea889929a68ccc1a2406ae3771d7b5ea9af14caaec52d943991404e0\": container with ID starting with e04f1d97ea889929a68ccc1a2406ae3771d7b5ea9af14caaec52d943991404e0 not found: ID does not exist" containerID="e04f1d97ea889929a68ccc1a2406ae3771d7b5ea9af14caaec52d943991404e0" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.902553 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e04f1d97ea889929a68ccc1a2406ae3771d7b5ea9af14caaec52d943991404e0"} err="failed to get container status \"e04f1d97ea889929a68ccc1a2406ae3771d7b5ea9af14caaec52d943991404e0\": rpc error: code = NotFound desc = could not find container \"e04f1d97ea889929a68ccc1a2406ae3771d7b5ea9af14caaec52d943991404e0\": container with ID starting with e04f1d97ea889929a68ccc1a2406ae3771d7b5ea9af14caaec52d943991404e0 not found: ID does not exist" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.902570 4948 scope.go:117] "RemoveContainer" containerID="43e2a82447778e081804b3e50b6f485e50305ec4eb1bbe2f231f259d11b87542" Feb 20 08:25:36 crc kubenswrapper[4948]: E0220 08:25:36.903018 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43e2a82447778e081804b3e50b6f485e50305ec4eb1bbe2f231f259d11b87542\": container with ID starting with 43e2a82447778e081804b3e50b6f485e50305ec4eb1bbe2f231f259d11b87542 not found: ID does not exist" containerID="43e2a82447778e081804b3e50b6f485e50305ec4eb1bbe2f231f259d11b87542" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.903045 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43e2a82447778e081804b3e50b6f485e50305ec4eb1bbe2f231f259d11b87542"} err="failed to get container status \"43e2a82447778e081804b3e50b6f485e50305ec4eb1bbe2f231f259d11b87542\": rpc error: code = NotFound desc = could not find container \"43e2a82447778e081804b3e50b6f485e50305ec4eb1bbe2f231f259d11b87542\": container with ID starting with 43e2a82447778e081804b3e50b6f485e50305ec4eb1bbe2f231f259d11b87542 not found: ID does not exist" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.903061 4948 scope.go:117] "RemoveContainer" containerID="d03b827c5401b1d3755870ce6ce51a7201888c77eda8df6a4497ae437d66e88a" Feb 20 08:25:36 crc kubenswrapper[4948]: E0220 08:25:36.903387 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d03b827c5401b1d3755870ce6ce51a7201888c77eda8df6a4497ae437d66e88a\": container with ID starting with d03b827c5401b1d3755870ce6ce51a7201888c77eda8df6a4497ae437d66e88a not found: ID does not exist" containerID="d03b827c5401b1d3755870ce6ce51a7201888c77eda8df6a4497ae437d66e88a" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.903428 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d03b827c5401b1d3755870ce6ce51a7201888c77eda8df6a4497ae437d66e88a"} err="failed to get container status \"d03b827c5401b1d3755870ce6ce51a7201888c77eda8df6a4497ae437d66e88a\": rpc error: code = NotFound desc = could not find container \"d03b827c5401b1d3755870ce6ce51a7201888c77eda8df6a4497ae437d66e88a\": container with ID starting with d03b827c5401b1d3755870ce6ce51a7201888c77eda8df6a4497ae437d66e88a not found: ID does not exist" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.927011 4948 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.927036 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.927046 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwwjr\" (UniqueName: \"kubernetes.io/projected/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-kube-api-access-mwwjr\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.927054 4948 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.936146 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-config-data" (OuterVolumeSpecName: "config-data") pod "be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" (UID: "be0e09bc-edd8-4ea9-8a59-ab630ed91cd8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.946121 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" (UID: "be0e09bc-edd8-4ea9-8a59-ab630ed91cd8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.029088 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.029124 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.097821 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.106637 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.129156 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:25:37 crc kubenswrapper[4948]: E0220 08:25:37.129611 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerName="proxy-httpd" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.129635 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerName="proxy-httpd" Feb 20 08:25:37 crc kubenswrapper[4948]: E0220 08:25:37.129666 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerName="sg-core" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.129676 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerName="sg-core" Feb 20 08:25:37 crc kubenswrapper[4948]: E0220 08:25:37.129684 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerName="ceilometer-notification-agent" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.129692 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerName="ceilometer-notification-agent" Feb 20 08:25:37 crc kubenswrapper[4948]: E0220 08:25:37.129723 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerName="ceilometer-central-agent" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.129731 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerName="ceilometer-central-agent" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.129930 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerName="ceilometer-notification-agent" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.129953 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerName="sg-core" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.129991 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerName="ceilometer-central-agent" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.130012 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerName="proxy-httpd" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.132039 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.136919 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.137783 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.138047 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.149200 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.231734 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58f563bb-9716-4abd-b187-3f01036d4f31-log-httpd\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.231818 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58f563bb-9716-4abd-b187-3f01036d4f31-run-httpd\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.231851 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58f563bb-9716-4abd-b187-3f01036d4f31-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.231888 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58f563bb-9716-4abd-b187-3f01036d4f31-config-data\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.231907 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/58f563bb-9716-4abd-b187-3f01036d4f31-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.231936 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xj8mt\" (UniqueName: \"kubernetes.io/projected/58f563bb-9716-4abd-b187-3f01036d4f31-kube-api-access-xj8mt\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.232004 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58f563bb-9716-4abd-b187-3f01036d4f31-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.232031 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58f563bb-9716-4abd-b187-3f01036d4f31-scripts\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.333603 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58f563bb-9716-4abd-b187-3f01036d4f31-scripts\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.333733 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58f563bb-9716-4abd-b187-3f01036d4f31-log-httpd\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.333785 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58f563bb-9716-4abd-b187-3f01036d4f31-run-httpd\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.333825 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58f563bb-9716-4abd-b187-3f01036d4f31-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.333871 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58f563bb-9716-4abd-b187-3f01036d4f31-config-data\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.333892 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/58f563bb-9716-4abd-b187-3f01036d4f31-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.333922 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xj8mt\" (UniqueName: \"kubernetes.io/projected/58f563bb-9716-4abd-b187-3f01036d4f31-kube-api-access-xj8mt\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.333999 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58f563bb-9716-4abd-b187-3f01036d4f31-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.335057 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58f563bb-9716-4abd-b187-3f01036d4f31-run-httpd\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.335637 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58f563bb-9716-4abd-b187-3f01036d4f31-log-httpd\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.339000 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/58f563bb-9716-4abd-b187-3f01036d4f31-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.341666 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58f563bb-9716-4abd-b187-3f01036d4f31-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.343083 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58f563bb-9716-4abd-b187-3f01036d4f31-scripts\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.347377 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58f563bb-9716-4abd-b187-3f01036d4f31-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.348839 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58f563bb-9716-4abd-b187-3f01036d4f31-config-data\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.362243 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xj8mt\" (UniqueName: \"kubernetes.io/projected/58f563bb-9716-4abd-b187-3f01036d4f31-kube-api-access-xj8mt\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.445818 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.733512 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" path="/var/lib/kubelet/pods/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8/volumes" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.809873 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:25:37 crc kubenswrapper[4948]: W0220 08:25:37.814395 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58f563bb_9716_4abd_b187_3f01036d4f31.slice/crio-f2e44f0bba52f19eac0a54be4d80b29f6c40e12f4cd9b659958bfeb7f47736f8 WatchSource:0}: Error finding container f2e44f0bba52f19eac0a54be4d80b29f6c40e12f4cd9b659958bfeb7f47736f8: Status 404 returned error can't find the container with id f2e44f0bba52f19eac0a54be4d80b29f6c40e12f4cd9b659958bfeb7f47736f8 Feb 20 08:25:38 crc kubenswrapper[4948]: I0220 08:25:38.800034 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58f563bb-9716-4abd-b187-3f01036d4f31","Type":"ContainerStarted","Data":"48565919319b61cf40009b1fc3e376ca0610ca1b8cffb64ba96185c28a75764b"} Feb 20 08:25:38 crc kubenswrapper[4948]: I0220 08:25:38.800446 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58f563bb-9716-4abd-b187-3f01036d4f31","Type":"ContainerStarted","Data":"f2e44f0bba52f19eac0a54be4d80b29f6c40e12f4cd9b659958bfeb7f47736f8"} Feb 20 08:25:39 crc kubenswrapper[4948]: I0220 08:25:39.231620 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 20 08:25:39 crc kubenswrapper[4948]: I0220 08:25:39.233216 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 20 08:25:39 crc kubenswrapper[4948]: I0220 08:25:39.240355 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 20 08:25:39 crc kubenswrapper[4948]: I0220 08:25:39.281490 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 20 08:25:39 crc kubenswrapper[4948]: I0220 08:25:39.285168 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 20 08:25:39 crc kubenswrapper[4948]: I0220 08:25:39.292165 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 20 08:25:39 crc kubenswrapper[4948]: I0220 08:25:39.300233 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 20 08:25:39 crc kubenswrapper[4948]: I0220 08:25:39.812493 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58f563bb-9716-4abd-b187-3f01036d4f31","Type":"ContainerStarted","Data":"9d613f1e4dac4655d24f385e90b54a96a440e203bbaebef9c0b52d894937a29e"} Feb 20 08:25:39 crc kubenswrapper[4948]: I0220 08:25:39.813233 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 20 08:25:39 crc kubenswrapper[4948]: I0220 08:25:39.816535 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 20 08:25:39 crc kubenswrapper[4948]: I0220 08:25:39.820938 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 20 08:25:40 crc kubenswrapper[4948]: I0220 08:25:40.824385 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58f563bb-9716-4abd-b187-3f01036d4f31","Type":"ContainerStarted","Data":"94b2bd264871b0e4898e842ab549234e35ff3a87346db7a895e5a9b2b8fc5897"} Feb 20 08:25:41 crc kubenswrapper[4948]: I0220 08:25:41.836459 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58f563bb-9716-4abd-b187-3f01036d4f31","Type":"ContainerStarted","Data":"3bcc9d21f8f1da159817d123e65bc83691e3ce0ff5eff3f6c34f4a99d28e9433"} Feb 20 08:25:41 crc kubenswrapper[4948]: I0220 08:25:41.867891 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.321293442 podStartE2EDuration="4.867864228s" podCreationTimestamp="2026-02-20 08:25:37 +0000 UTC" firstStartedPulling="2026-02-20 08:25:37.817180576 +0000 UTC m=+1186.791675416" lastFinishedPulling="2026-02-20 08:25:41.363751382 +0000 UTC m=+1190.338246202" observedRunningTime="2026-02-20 08:25:41.85634308 +0000 UTC m=+1190.830837940" watchObservedRunningTime="2026-02-20 08:25:41.867864228 +0000 UTC m=+1190.842359088" Feb 20 08:25:42 crc kubenswrapper[4948]: I0220 08:25:42.854258 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 20 08:25:44 crc kubenswrapper[4948]: I0220 08:25:44.134112 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 20 08:26:07 crc kubenswrapper[4948]: I0220 08:26:07.454550 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 20 08:26:16 crc kubenswrapper[4948]: I0220 08:26:16.577316 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 20 08:26:17 crc kubenswrapper[4948]: I0220 08:26:17.598386 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 20 08:26:20 crc kubenswrapper[4948]: I0220 08:26:20.589002 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="73b84bb7-f594-4823-ac03-40fdac6ee177" containerName="rabbitmq" containerID="cri-o://f2ef52eaf8fede6026f04fb9c9c143db4ac26c52cfdab80a2a52811b838b1e85" gracePeriod=604796 Feb 20 08:26:21 crc kubenswrapper[4948]: I0220 08:26:21.523688 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="790894c9-053c-497c-955d-ce7519111dd6" containerName="rabbitmq" containerID="cri-o://ba12ca5de96a23b6d0de36b28eae9e3fb2a92d4464edd4582c10c73ae804254a" gracePeriod=604797 Feb 20 08:26:24 crc kubenswrapper[4948]: I0220 08:26:24.117591 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="73b84bb7-f594-4823-ac03-40fdac6ee177" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.94:5671: connect: connection refused" Feb 20 08:26:24 crc kubenswrapper[4948]: I0220 08:26:24.498402 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="790894c9-053c-497c-955d-ce7519111dd6" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.95:5671: connect: connection refused" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.289632 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.349154 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/73b84bb7-f594-4823-ac03-40fdac6ee177-pod-info\") pod \"73b84bb7-f594-4823-ac03-40fdac6ee177\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.349228 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-server-conf\") pod \"73b84bb7-f594-4823-ac03-40fdac6ee177\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.349244 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-tls\") pod \"73b84bb7-f594-4823-ac03-40fdac6ee177\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.349270 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-config-data\") pod \"73b84bb7-f594-4823-ac03-40fdac6ee177\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.349301 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-erlang-cookie\") pod \"73b84bb7-f594-4823-ac03-40fdac6ee177\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.349347 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"73b84bb7-f594-4823-ac03-40fdac6ee177\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.349380 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/73b84bb7-f594-4823-ac03-40fdac6ee177-erlang-cookie-secret\") pod \"73b84bb7-f594-4823-ac03-40fdac6ee177\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.349428 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-plugins\") pod \"73b84bb7-f594-4823-ac03-40fdac6ee177\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.349471 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-plugins-conf\") pod \"73b84bb7-f594-4823-ac03-40fdac6ee177\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.349490 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-confd\") pod \"73b84bb7-f594-4823-ac03-40fdac6ee177\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.349556 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqlcq\" (UniqueName: \"kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-kube-api-access-tqlcq\") pod \"73b84bb7-f594-4823-ac03-40fdac6ee177\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.350426 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "73b84bb7-f594-4823-ac03-40fdac6ee177" (UID: "73b84bb7-f594-4823-ac03-40fdac6ee177"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.350628 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "73b84bb7-f594-4823-ac03-40fdac6ee177" (UID: "73b84bb7-f594-4823-ac03-40fdac6ee177"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.350733 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "73b84bb7-f594-4823-ac03-40fdac6ee177" (UID: "73b84bb7-f594-4823-ac03-40fdac6ee177"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.356708 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-kube-api-access-tqlcq" (OuterVolumeSpecName: "kube-api-access-tqlcq") pod "73b84bb7-f594-4823-ac03-40fdac6ee177" (UID: "73b84bb7-f594-4823-ac03-40fdac6ee177"). InnerVolumeSpecName "kube-api-access-tqlcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.358193 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/73b84bb7-f594-4823-ac03-40fdac6ee177-pod-info" (OuterVolumeSpecName: "pod-info") pod "73b84bb7-f594-4823-ac03-40fdac6ee177" (UID: "73b84bb7-f594-4823-ac03-40fdac6ee177"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.358424 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73b84bb7-f594-4823-ac03-40fdac6ee177-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "73b84bb7-f594-4823-ac03-40fdac6ee177" (UID: "73b84bb7-f594-4823-ac03-40fdac6ee177"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.367317 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "73b84bb7-f594-4823-ac03-40fdac6ee177" (UID: "73b84bb7-f594-4823-ac03-40fdac6ee177"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.374196 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "persistence") pod "73b84bb7-f594-4823-ac03-40fdac6ee177" (UID: "73b84bb7-f594-4823-ac03-40fdac6ee177"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.415847 4948 generic.go:334] "Generic (PLEG): container finished" podID="73b84bb7-f594-4823-ac03-40fdac6ee177" containerID="f2ef52eaf8fede6026f04fb9c9c143db4ac26c52cfdab80a2a52811b838b1e85" exitCode=0 Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.415891 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"73b84bb7-f594-4823-ac03-40fdac6ee177","Type":"ContainerDied","Data":"f2ef52eaf8fede6026f04fb9c9c143db4ac26c52cfdab80a2a52811b838b1e85"} Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.415916 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"73b84bb7-f594-4823-ac03-40fdac6ee177","Type":"ContainerDied","Data":"e552b8693d7f82c1ed5a2c7c103ec119814c444b37d5f93e2491692378454a4c"} Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.415932 4948 scope.go:117] "RemoveContainer" containerID="f2ef52eaf8fede6026f04fb9c9c143db4ac26c52cfdab80a2a52811b838b1e85" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.416796 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.427239 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-config-data" (OuterVolumeSpecName: "config-data") pod "73b84bb7-f594-4823-ac03-40fdac6ee177" (UID: "73b84bb7-f594-4823-ac03-40fdac6ee177"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.452577 4948 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/73b84bb7-f594-4823-ac03-40fdac6ee177-pod-info\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.452609 4948 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.452624 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.452636 4948 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.452682 4948 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.452694 4948 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/73b84bb7-f594-4823-ac03-40fdac6ee177-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.452708 4948 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.459045 4948 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.459095 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqlcq\" (UniqueName: \"kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-kube-api-access-tqlcq\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.476131 4948 scope.go:117] "RemoveContainer" containerID="d4b1f06d6e332e362b758285a5271800ff6a5cc95aee58142ef4093fc832e364" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.477161 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-server-conf" (OuterVolumeSpecName: "server-conf") pod "73b84bb7-f594-4823-ac03-40fdac6ee177" (UID: "73b84bb7-f594-4823-ac03-40fdac6ee177"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.501147 4948 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.514873 4948 scope.go:117] "RemoveContainer" containerID="f2ef52eaf8fede6026f04fb9c9c143db4ac26c52cfdab80a2a52811b838b1e85" Feb 20 08:26:27 crc kubenswrapper[4948]: E0220 08:26:27.515316 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2ef52eaf8fede6026f04fb9c9c143db4ac26c52cfdab80a2a52811b838b1e85\": container with ID starting with f2ef52eaf8fede6026f04fb9c9c143db4ac26c52cfdab80a2a52811b838b1e85 not found: ID does not exist" containerID="f2ef52eaf8fede6026f04fb9c9c143db4ac26c52cfdab80a2a52811b838b1e85" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.515344 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2ef52eaf8fede6026f04fb9c9c143db4ac26c52cfdab80a2a52811b838b1e85"} err="failed to get container status \"f2ef52eaf8fede6026f04fb9c9c143db4ac26c52cfdab80a2a52811b838b1e85\": rpc error: code = NotFound desc = could not find container \"f2ef52eaf8fede6026f04fb9c9c143db4ac26c52cfdab80a2a52811b838b1e85\": container with ID starting with f2ef52eaf8fede6026f04fb9c9c143db4ac26c52cfdab80a2a52811b838b1e85 not found: ID does not exist" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.515362 4948 scope.go:117] "RemoveContainer" containerID="d4b1f06d6e332e362b758285a5271800ff6a5cc95aee58142ef4093fc832e364" Feb 20 08:26:27 crc kubenswrapper[4948]: E0220 08:26:27.515696 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4b1f06d6e332e362b758285a5271800ff6a5cc95aee58142ef4093fc832e364\": container with ID starting with d4b1f06d6e332e362b758285a5271800ff6a5cc95aee58142ef4093fc832e364 not found: ID does not exist" containerID="d4b1f06d6e332e362b758285a5271800ff6a5cc95aee58142ef4093fc832e364" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.515716 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4b1f06d6e332e362b758285a5271800ff6a5cc95aee58142ef4093fc832e364"} err="failed to get container status \"d4b1f06d6e332e362b758285a5271800ff6a5cc95aee58142ef4093fc832e364\": rpc error: code = NotFound desc = could not find container \"d4b1f06d6e332e362b758285a5271800ff6a5cc95aee58142ef4093fc832e364\": container with ID starting with d4b1f06d6e332e362b758285a5271800ff6a5cc95aee58142ef4093fc832e364 not found: ID does not exist" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.551504 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "73b84bb7-f594-4823-ac03-40fdac6ee177" (UID: "73b84bb7-f594-4823-ac03-40fdac6ee177"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.561620 4948 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-server-conf\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.562009 4948 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.562083 4948 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.766423 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.787412 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.804985 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 20 08:26:27 crc kubenswrapper[4948]: E0220 08:26:27.805629 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73b84bb7-f594-4823-ac03-40fdac6ee177" containerName="rabbitmq" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.805768 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="73b84bb7-f594-4823-ac03-40fdac6ee177" containerName="rabbitmq" Feb 20 08:26:27 crc kubenswrapper[4948]: E0220 08:26:27.805860 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73b84bb7-f594-4823-ac03-40fdac6ee177" containerName="setup-container" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.805941 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="73b84bb7-f594-4823-ac03-40fdac6ee177" containerName="setup-container" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.806267 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="73b84bb7-f594-4823-ac03-40fdac6ee177" containerName="rabbitmq" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.807649 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.811538 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.811834 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.826253 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.827283 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.827486 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.827598 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.827696 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.827795 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-8zcmc" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.868096 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f063bb62-10c3-46d9-a41d-56d2826059e0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.868141 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjqdv\" (UniqueName: \"kubernetes.io/projected/f063bb62-10c3-46d9-a41d-56d2826059e0-kube-api-access-mjqdv\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.868163 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f063bb62-10c3-46d9-a41d-56d2826059e0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.868201 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f063bb62-10c3-46d9-a41d-56d2826059e0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.868225 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f063bb62-10c3-46d9-a41d-56d2826059e0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.868241 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f063bb62-10c3-46d9-a41d-56d2826059e0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.868331 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f063bb62-10c3-46d9-a41d-56d2826059e0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.868371 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.868411 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f063bb62-10c3-46d9-a41d-56d2826059e0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.868441 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f063bb62-10c3-46d9-a41d-56d2826059e0-config-data\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.868482 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f063bb62-10c3-46d9-a41d-56d2826059e0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.970191 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.970274 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f063bb62-10c3-46d9-a41d-56d2826059e0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.970328 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f063bb62-10c3-46d9-a41d-56d2826059e0-config-data\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.970388 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f063bb62-10c3-46d9-a41d-56d2826059e0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.970427 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f063bb62-10c3-46d9-a41d-56d2826059e0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.970467 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjqdv\" (UniqueName: \"kubernetes.io/projected/f063bb62-10c3-46d9-a41d-56d2826059e0-kube-api-access-mjqdv\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.970485 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f063bb62-10c3-46d9-a41d-56d2826059e0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.970513 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f063bb62-10c3-46d9-a41d-56d2826059e0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.970543 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.972137 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f063bb62-10c3-46d9-a41d-56d2826059e0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.970553 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f063bb62-10c3-46d9-a41d-56d2826059e0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.978485 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f063bb62-10c3-46d9-a41d-56d2826059e0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.978577 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f063bb62-10c3-46d9-a41d-56d2826059e0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.979549 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f063bb62-10c3-46d9-a41d-56d2826059e0-config-data\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.980052 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f063bb62-10c3-46d9-a41d-56d2826059e0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.980416 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f063bb62-10c3-46d9-a41d-56d2826059e0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.980520 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f063bb62-10c3-46d9-a41d-56d2826059e0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.983102 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f063bb62-10c3-46d9-a41d-56d2826059e0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.985995 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f063bb62-10c3-46d9-a41d-56d2826059e0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.992329 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f063bb62-10c3-46d9-a41d-56d2826059e0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.995558 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f063bb62-10c3-46d9-a41d-56d2826059e0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.002144 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjqdv\" (UniqueName: \"kubernetes.io/projected/f063bb62-10c3-46d9-a41d-56d2826059e0-kube-api-access-mjqdv\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.032669 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.127685 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.193813 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.282662 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-plugins-conf\") pod \"790894c9-053c-497c-955d-ce7519111dd6\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.282894 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-plugins\") pod \"790894c9-053c-497c-955d-ce7519111dd6\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.283024 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7t2f\" (UniqueName: \"kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-kube-api-access-b7t2f\") pod \"790894c9-053c-497c-955d-ce7519111dd6\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.283085 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-tls\") pod \"790894c9-053c-497c-955d-ce7519111dd6\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.283113 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/790894c9-053c-497c-955d-ce7519111dd6-erlang-cookie-secret\") pod \"790894c9-053c-497c-955d-ce7519111dd6\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.283164 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-confd\") pod \"790894c9-053c-497c-955d-ce7519111dd6\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.283211 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-server-conf\") pod \"790894c9-053c-497c-955d-ce7519111dd6\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.283235 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"790894c9-053c-497c-955d-ce7519111dd6\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.283277 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/790894c9-053c-497c-955d-ce7519111dd6-pod-info\") pod \"790894c9-053c-497c-955d-ce7519111dd6\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.283309 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-erlang-cookie\") pod \"790894c9-053c-497c-955d-ce7519111dd6\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.283338 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-config-data\") pod \"790894c9-053c-497c-955d-ce7519111dd6\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.283548 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "790894c9-053c-497c-955d-ce7519111dd6" (UID: "790894c9-053c-497c-955d-ce7519111dd6"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.283863 4948 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.283863 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "790894c9-053c-497c-955d-ce7519111dd6" (UID: "790894c9-053c-497c-955d-ce7519111dd6"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.289690 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "790894c9-053c-497c-955d-ce7519111dd6" (UID: "790894c9-053c-497c-955d-ce7519111dd6"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.295770 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "790894c9-053c-497c-955d-ce7519111dd6" (UID: "790894c9-053c-497c-955d-ce7519111dd6"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.295785 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/790894c9-053c-497c-955d-ce7519111dd6-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "790894c9-053c-497c-955d-ce7519111dd6" (UID: "790894c9-053c-497c-955d-ce7519111dd6"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.296304 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "790894c9-053c-497c-955d-ce7519111dd6" (UID: "790894c9-053c-497c-955d-ce7519111dd6"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.298357 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-kube-api-access-b7t2f" (OuterVolumeSpecName: "kube-api-access-b7t2f") pod "790894c9-053c-497c-955d-ce7519111dd6" (UID: "790894c9-053c-497c-955d-ce7519111dd6"). InnerVolumeSpecName "kube-api-access-b7t2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.299100 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/790894c9-053c-497c-955d-ce7519111dd6-pod-info" (OuterVolumeSpecName: "pod-info") pod "790894c9-053c-497c-955d-ce7519111dd6" (UID: "790894c9-053c-497c-955d-ce7519111dd6"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.343736 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-config-data" (OuterVolumeSpecName: "config-data") pod "790894c9-053c-497c-955d-ce7519111dd6" (UID: "790894c9-053c-497c-955d-ce7519111dd6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.356996 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-server-conf" (OuterVolumeSpecName: "server-conf") pod "790894c9-053c-497c-955d-ce7519111dd6" (UID: "790894c9-053c-497c-955d-ce7519111dd6"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.385712 4948 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.385742 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7t2f\" (UniqueName: \"kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-kube-api-access-b7t2f\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.385752 4948 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.385763 4948 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/790894c9-053c-497c-955d-ce7519111dd6-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.385771 4948 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-server-conf\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.385790 4948 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.385800 4948 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/790894c9-053c-497c-955d-ce7519111dd6-pod-info\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.385808 4948 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.385816 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.409318 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "790894c9-053c-497c-955d-ce7519111dd6" (UID: "790894c9-053c-497c-955d-ce7519111dd6"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.415366 4948 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.425897 4948 generic.go:334] "Generic (PLEG): container finished" podID="790894c9-053c-497c-955d-ce7519111dd6" containerID="ba12ca5de96a23b6d0de36b28eae9e3fb2a92d4464edd4582c10c73ae804254a" exitCode=0 Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.425961 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"790894c9-053c-497c-955d-ce7519111dd6","Type":"ContainerDied","Data":"ba12ca5de96a23b6d0de36b28eae9e3fb2a92d4464edd4582c10c73ae804254a"} Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.426001 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.426065 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"790894c9-053c-497c-955d-ce7519111dd6","Type":"ContainerDied","Data":"39af94a050f39ff7ebfa915fc855416f40bbca9d8ae91ba23049ec9be5e37b76"} Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.426103 4948 scope.go:117] "RemoveContainer" containerID="ba12ca5de96a23b6d0de36b28eae9e3fb2a92d4464edd4582c10c73ae804254a" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.458416 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.472054 4948 scope.go:117] "RemoveContainer" containerID="77b026b61409d73dee6a89b7f49bf1a86167df7972db0e3eb3f4cbb10df48242" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.478122 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.487261 4948 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.487290 4948 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.494818 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 20 08:26:28 crc kubenswrapper[4948]: E0220 08:26:28.495207 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="790894c9-053c-497c-955d-ce7519111dd6" containerName="setup-container" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.495227 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="790894c9-053c-497c-955d-ce7519111dd6" containerName="setup-container" Feb 20 08:26:28 crc kubenswrapper[4948]: E0220 08:26:28.495252 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="790894c9-053c-497c-955d-ce7519111dd6" containerName="rabbitmq" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.495259 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="790894c9-053c-497c-955d-ce7519111dd6" containerName="rabbitmq" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.495439 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="790894c9-053c-497c-955d-ce7519111dd6" containerName="rabbitmq" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.496303 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.501695 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.501713 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.501875 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.502299 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.502395 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.502476 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-rrjt5" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.502919 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.515070 4948 scope.go:117] "RemoveContainer" containerID="ba12ca5de96a23b6d0de36b28eae9e3fb2a92d4464edd4582c10c73ae804254a" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.515585 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 20 08:26:28 crc kubenswrapper[4948]: E0220 08:26:28.520793 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba12ca5de96a23b6d0de36b28eae9e3fb2a92d4464edd4582c10c73ae804254a\": container with ID starting with ba12ca5de96a23b6d0de36b28eae9e3fb2a92d4464edd4582c10c73ae804254a not found: ID does not exist" containerID="ba12ca5de96a23b6d0de36b28eae9e3fb2a92d4464edd4582c10c73ae804254a" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.520827 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba12ca5de96a23b6d0de36b28eae9e3fb2a92d4464edd4582c10c73ae804254a"} err="failed to get container status \"ba12ca5de96a23b6d0de36b28eae9e3fb2a92d4464edd4582c10c73ae804254a\": rpc error: code = NotFound desc = could not find container \"ba12ca5de96a23b6d0de36b28eae9e3fb2a92d4464edd4582c10c73ae804254a\": container with ID starting with ba12ca5de96a23b6d0de36b28eae9e3fb2a92d4464edd4582c10c73ae804254a not found: ID does not exist" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.520848 4948 scope.go:117] "RemoveContainer" containerID="77b026b61409d73dee6a89b7f49bf1a86167df7972db0e3eb3f4cbb10df48242" Feb 20 08:26:28 crc kubenswrapper[4948]: E0220 08:26:28.527463 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77b026b61409d73dee6a89b7f49bf1a86167df7972db0e3eb3f4cbb10df48242\": container with ID starting with 77b026b61409d73dee6a89b7f49bf1a86167df7972db0e3eb3f4cbb10df48242 not found: ID does not exist" containerID="77b026b61409d73dee6a89b7f49bf1a86167df7972db0e3eb3f4cbb10df48242" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.527650 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77b026b61409d73dee6a89b7f49bf1a86167df7972db0e3eb3f4cbb10df48242"} err="failed to get container status \"77b026b61409d73dee6a89b7f49bf1a86167df7972db0e3eb3f4cbb10df48242\": rpc error: code = NotFound desc = could not find container \"77b026b61409d73dee6a89b7f49bf1a86167df7972db0e3eb3f4cbb10df48242\": container with ID starting with 77b026b61409d73dee6a89b7f49bf1a86167df7972db0e3eb3f4cbb10df48242 not found: ID does not exist" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.588289 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/037b329f-712f-4a67-984a-75affd2a57b7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.588341 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/037b329f-712f-4a67-984a-75affd2a57b7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.588383 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/037b329f-712f-4a67-984a-75affd2a57b7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.588425 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fn7kw\" (UniqueName: \"kubernetes.io/projected/037b329f-712f-4a67-984a-75affd2a57b7-kube-api-access-fn7kw\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.588452 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/037b329f-712f-4a67-984a-75affd2a57b7-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.588471 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.588494 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/037b329f-712f-4a67-984a-75affd2a57b7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.588514 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/037b329f-712f-4a67-984a-75affd2a57b7-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.588559 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/037b329f-712f-4a67-984a-75affd2a57b7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.588573 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/037b329f-712f-4a67-984a-75affd2a57b7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.588594 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/037b329f-712f-4a67-984a-75affd2a57b7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.651421 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.690347 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/037b329f-712f-4a67-984a-75affd2a57b7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.690399 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/037b329f-712f-4a67-984a-75affd2a57b7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.690440 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/037b329f-712f-4a67-984a-75affd2a57b7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.690463 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/037b329f-712f-4a67-984a-75affd2a57b7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.690499 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/037b329f-712f-4a67-984a-75affd2a57b7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.690549 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/037b329f-712f-4a67-984a-75affd2a57b7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.690599 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fn7kw\" (UniqueName: \"kubernetes.io/projected/037b329f-712f-4a67-984a-75affd2a57b7-kube-api-access-fn7kw\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.690635 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/037b329f-712f-4a67-984a-75affd2a57b7-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.690663 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.690726 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/037b329f-712f-4a67-984a-75affd2a57b7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.690759 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/037b329f-712f-4a67-984a-75affd2a57b7-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.691297 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/037b329f-712f-4a67-984a-75affd2a57b7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.691437 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.692711 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/037b329f-712f-4a67-984a-75affd2a57b7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.693318 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/037b329f-712f-4a67-984a-75affd2a57b7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.693530 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/037b329f-712f-4a67-984a-75affd2a57b7-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.693667 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/037b329f-712f-4a67-984a-75affd2a57b7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.695055 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/037b329f-712f-4a67-984a-75affd2a57b7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.696171 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/037b329f-712f-4a67-984a-75affd2a57b7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.700467 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/037b329f-712f-4a67-984a-75affd2a57b7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.707460 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/037b329f-712f-4a67-984a-75affd2a57b7-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.710169 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fn7kw\" (UniqueName: \"kubernetes.io/projected/037b329f-712f-4a67-984a-75affd2a57b7-kube-api-access-fn7kw\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.736956 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.822113 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.166655 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-vgk45"] Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.169046 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.171304 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.193784 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-vgk45"] Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.200355 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.200421 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.200465 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-config\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.200565 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.200614 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.200658 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.200726 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4h8d\" (UniqueName: \"kubernetes.io/projected/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-kube-api-access-x4h8d\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.302075 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.302149 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4h8d\" (UniqueName: \"kubernetes.io/projected/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-kube-api-access-x4h8d\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.302194 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.302219 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.302256 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-config\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.302317 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.302349 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.303354 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.303881 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.304634 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.305259 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.305792 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-config\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.306018 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.306245 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.322835 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4h8d\" (UniqueName: \"kubernetes.io/projected/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-kube-api-access-x4h8d\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.444335 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"037b329f-712f-4a67-984a-75affd2a57b7","Type":"ContainerStarted","Data":"5523f651b79457ea6012384e30f4ce29cd629dcb583a6414256e8cccebe13e53"} Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.446479 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f063bb62-10c3-46d9-a41d-56d2826059e0","Type":"ContainerStarted","Data":"49966811dc211199039188fd9f5a17caac95e7a7a6cc21002f62432ed349328d"} Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.495117 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.752025 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73b84bb7-f594-4823-ac03-40fdac6ee177" path="/var/lib/kubelet/pods/73b84bb7-f594-4823-ac03-40fdac6ee177/volumes" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.754326 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="790894c9-053c-497c-955d-ce7519111dd6" path="/var/lib/kubelet/pods/790894c9-053c-497c-955d-ce7519111dd6/volumes" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.998387 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-vgk45"] Feb 20 08:26:29 crc kubenswrapper[4948]: W0220 08:26:29.999620 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb4bae750_8bf9_4804_bab7_9c94d4be7bb6.slice/crio-a5fc6b8697916867d170463ef9a68553f0374e86750fe3d5b67599ef15a68b91 WatchSource:0}: Error finding container a5fc6b8697916867d170463ef9a68553f0374e86750fe3d5b67599ef15a68b91: Status 404 returned error can't find the container with id a5fc6b8697916867d170463ef9a68553f0374e86750fe3d5b67599ef15a68b91 Feb 20 08:26:30 crc kubenswrapper[4948]: I0220 08:26:30.466074 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f063bb62-10c3-46d9-a41d-56d2826059e0","Type":"ContainerStarted","Data":"878f9968270125c43cab5cdbc68ca06efa6e152bfb9c605c96475afc50f8ad38"} Feb 20 08:26:30 crc kubenswrapper[4948]: I0220 08:26:30.470427 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" event={"ID":"b4bae750-8bf9-4804-bab7-9c94d4be7bb6","Type":"ContainerStarted","Data":"a5fc6b8697916867d170463ef9a68553f0374e86750fe3d5b67599ef15a68b91"} Feb 20 08:26:31 crc kubenswrapper[4948]: I0220 08:26:31.482959 4948 generic.go:334] "Generic (PLEG): container finished" podID="b4bae750-8bf9-4804-bab7-9c94d4be7bb6" containerID="6457842f65c42b8313a38ca922ee61a97a4823ed4c56040981289a5c5705cd46" exitCode=0 Feb 20 08:26:31 crc kubenswrapper[4948]: I0220 08:26:31.483042 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" event={"ID":"b4bae750-8bf9-4804-bab7-9c94d4be7bb6","Type":"ContainerDied","Data":"6457842f65c42b8313a38ca922ee61a97a4823ed4c56040981289a5c5705cd46"} Feb 20 08:26:31 crc kubenswrapper[4948]: I0220 08:26:31.485969 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"037b329f-712f-4a67-984a-75affd2a57b7","Type":"ContainerStarted","Data":"9fcd3d9744d41147b3cfe90374517e1e56f49253e12cf9785df1944e69f8e3a2"} Feb 20 08:26:32 crc kubenswrapper[4948]: I0220 08:26:32.506872 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" event={"ID":"b4bae750-8bf9-4804-bab7-9c94d4be7bb6","Type":"ContainerStarted","Data":"ede14cf7eee35beb81bccaf96c5abbd270f76ba98b65d67579bbb482b130f45e"} Feb 20 08:26:32 crc kubenswrapper[4948]: I0220 08:26:32.507355 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:32 crc kubenswrapper[4948]: I0220 08:26:32.545562 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" podStartSLOduration=3.545532165 podStartE2EDuration="3.545532165s" podCreationTimestamp="2026-02-20 08:26:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:26:32.536469842 +0000 UTC m=+1241.510964702" watchObservedRunningTime="2026-02-20 08:26:32.545532165 +0000 UTC m=+1241.520027015" Feb 20 08:26:38 crc kubenswrapper[4948]: I0220 08:26:38.024811 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:26:38 crc kubenswrapper[4948]: I0220 08:26:38.025429 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.497503 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.609662 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-t56xb"] Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.609948 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" podUID="c1847849-cdf4-4ea6-a540-c01503a003b2" containerName="dnsmasq-dns" containerID="cri-o://5c65f781c2289e22368c8df4db43181d77d0c56ecf63f298c154f0b8af27a14f" gracePeriod=10 Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.805445 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55478c4467-fbjpk"] Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.811756 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.820757 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55478c4467-fbjpk"] Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.831595 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.831638 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.831688 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.831750 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xksr2\" (UniqueName: \"kubernetes.io/projected/b05f92c6-60c9-41be-bf6c-6288cee04659-kube-api-access-xksr2\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.831793 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-config\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.831826 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-dns-svc\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.831899 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.933429 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.933483 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.933558 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.933617 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xksr2\" (UniqueName: \"kubernetes.io/projected/b05f92c6-60c9-41be-bf6c-6288cee04659-kube-api-access-xksr2\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.933667 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-config\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.933714 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-dns-svc\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.933737 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.934783 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.934844 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.934869 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-dns-svc\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.934890 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-config\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.936874 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.936907 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.967132 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xksr2\" (UniqueName: \"kubernetes.io/projected/b05f92c6-60c9-41be-bf6c-6288cee04659-kube-api-access-xksr2\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.130019 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.145345 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.342071 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-ovsdbserver-nb\") pod \"c1847849-cdf4-4ea6-a540-c01503a003b2\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.342500 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-dns-svc\") pod \"c1847849-cdf4-4ea6-a540-c01503a003b2\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.342551 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-config\") pod \"c1847849-cdf4-4ea6-a540-c01503a003b2\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.342703 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-db22s\" (UniqueName: \"kubernetes.io/projected/c1847849-cdf4-4ea6-a540-c01503a003b2-kube-api-access-db22s\") pod \"c1847849-cdf4-4ea6-a540-c01503a003b2\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.342739 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-ovsdbserver-sb\") pod \"c1847849-cdf4-4ea6-a540-c01503a003b2\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.343519 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-dns-swift-storage-0\") pod \"c1847849-cdf4-4ea6-a540-c01503a003b2\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.349092 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1847849-cdf4-4ea6-a540-c01503a003b2-kube-api-access-db22s" (OuterVolumeSpecName: "kube-api-access-db22s") pod "c1847849-cdf4-4ea6-a540-c01503a003b2" (UID: "c1847849-cdf4-4ea6-a540-c01503a003b2"). InnerVolumeSpecName "kube-api-access-db22s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.405292 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c1847849-cdf4-4ea6-a540-c01503a003b2" (UID: "c1847849-cdf4-4ea6-a540-c01503a003b2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.405571 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c1847849-cdf4-4ea6-a540-c01503a003b2" (UID: "c1847849-cdf4-4ea6-a540-c01503a003b2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.409160 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-config" (OuterVolumeSpecName: "config") pod "c1847849-cdf4-4ea6-a540-c01503a003b2" (UID: "c1847849-cdf4-4ea6-a540-c01503a003b2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.413766 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c1847849-cdf4-4ea6-a540-c01503a003b2" (UID: "c1847849-cdf4-4ea6-a540-c01503a003b2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.427498 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c1847849-cdf4-4ea6-a540-c01503a003b2" (UID: "c1847849-cdf4-4ea6-a540-c01503a003b2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.445637 4948 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.445684 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.445698 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-db22s\" (UniqueName: \"kubernetes.io/projected/c1847849-cdf4-4ea6-a540-c01503a003b2-kube-api-access-db22s\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.445713 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.445725 4948 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.445736 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.594191 4948 generic.go:334] "Generic (PLEG): container finished" podID="c1847849-cdf4-4ea6-a540-c01503a003b2" containerID="5c65f781c2289e22368c8df4db43181d77d0c56ecf63f298c154f0b8af27a14f" exitCode=0 Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.594238 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.594250 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" event={"ID":"c1847849-cdf4-4ea6-a540-c01503a003b2","Type":"ContainerDied","Data":"5c65f781c2289e22368c8df4db43181d77d0c56ecf63f298c154f0b8af27a14f"} Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.594293 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" event={"ID":"c1847849-cdf4-4ea6-a540-c01503a003b2","Type":"ContainerDied","Data":"4b70d018879f9b501eed3beab4a8c0781964989ca45e86dd4327f865ae78f6bf"} Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.594315 4948 scope.go:117] "RemoveContainer" containerID="5c65f781c2289e22368c8df4db43181d77d0c56ecf63f298c154f0b8af27a14f" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.597965 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55478c4467-fbjpk"] Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.631714 4948 scope.go:117] "RemoveContainer" containerID="6b70c340bd4972b75cf8aa94cef8eb5d658d1c6345ae3959074d416c3683e47a" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.634751 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-t56xb"] Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.645136 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-t56xb"] Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.666425 4948 scope.go:117] "RemoveContainer" containerID="5c65f781c2289e22368c8df4db43181d77d0c56ecf63f298c154f0b8af27a14f" Feb 20 08:26:40 crc kubenswrapper[4948]: E0220 08:26:40.666876 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c65f781c2289e22368c8df4db43181d77d0c56ecf63f298c154f0b8af27a14f\": container with ID starting with 5c65f781c2289e22368c8df4db43181d77d0c56ecf63f298c154f0b8af27a14f not found: ID does not exist" containerID="5c65f781c2289e22368c8df4db43181d77d0c56ecf63f298c154f0b8af27a14f" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.667000 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c65f781c2289e22368c8df4db43181d77d0c56ecf63f298c154f0b8af27a14f"} err="failed to get container status \"5c65f781c2289e22368c8df4db43181d77d0c56ecf63f298c154f0b8af27a14f\": rpc error: code = NotFound desc = could not find container \"5c65f781c2289e22368c8df4db43181d77d0c56ecf63f298c154f0b8af27a14f\": container with ID starting with 5c65f781c2289e22368c8df4db43181d77d0c56ecf63f298c154f0b8af27a14f not found: ID does not exist" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.667090 4948 scope.go:117] "RemoveContainer" containerID="6b70c340bd4972b75cf8aa94cef8eb5d658d1c6345ae3959074d416c3683e47a" Feb 20 08:26:40 crc kubenswrapper[4948]: E0220 08:26:40.667540 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b70c340bd4972b75cf8aa94cef8eb5d658d1c6345ae3959074d416c3683e47a\": container with ID starting with 6b70c340bd4972b75cf8aa94cef8eb5d658d1c6345ae3959074d416c3683e47a not found: ID does not exist" containerID="6b70c340bd4972b75cf8aa94cef8eb5d658d1c6345ae3959074d416c3683e47a" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.668170 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b70c340bd4972b75cf8aa94cef8eb5d658d1c6345ae3959074d416c3683e47a"} err="failed to get container status \"6b70c340bd4972b75cf8aa94cef8eb5d658d1c6345ae3959074d416c3683e47a\": rpc error: code = NotFound desc = could not find container \"6b70c340bd4972b75cf8aa94cef8eb5d658d1c6345ae3959074d416c3683e47a\": container with ID starting with 6b70c340bd4972b75cf8aa94cef8eb5d658d1c6345ae3959074d416c3683e47a not found: ID does not exist" Feb 20 08:26:41 crc kubenswrapper[4948]: I0220 08:26:41.605352 4948 generic.go:334] "Generic (PLEG): container finished" podID="b05f92c6-60c9-41be-bf6c-6288cee04659" containerID="bdea951a4170c0cdceb0ddcae984666862ac3f74ed32191e1024d762cf306aa7" exitCode=0 Feb 20 08:26:41 crc kubenswrapper[4948]: I0220 08:26:41.605451 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-fbjpk" event={"ID":"b05f92c6-60c9-41be-bf6c-6288cee04659","Type":"ContainerDied","Data":"bdea951a4170c0cdceb0ddcae984666862ac3f74ed32191e1024d762cf306aa7"} Feb 20 08:26:41 crc kubenswrapper[4948]: I0220 08:26:41.607682 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-fbjpk" event={"ID":"b05f92c6-60c9-41be-bf6c-6288cee04659","Type":"ContainerStarted","Data":"8881488afcebfb5387b7d34be5b2ee2caf4be9af99e858b4f8b197f620f5b16f"} Feb 20 08:26:41 crc kubenswrapper[4948]: I0220 08:26:41.735406 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1847849-cdf4-4ea6-a540-c01503a003b2" path="/var/lib/kubelet/pods/c1847849-cdf4-4ea6-a540-c01503a003b2/volumes" Feb 20 08:26:42 crc kubenswrapper[4948]: I0220 08:26:42.630865 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-fbjpk" event={"ID":"b05f92c6-60c9-41be-bf6c-6288cee04659","Type":"ContainerStarted","Data":"180d9e61aab4c578242860dafa196fb5407f99775735708fe0b437e89f49320a"} Feb 20 08:26:42 crc kubenswrapper[4948]: I0220 08:26:42.631060 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:42 crc kubenswrapper[4948]: I0220 08:26:42.679102 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55478c4467-fbjpk" podStartSLOduration=3.679075083 podStartE2EDuration="3.679075083s" podCreationTimestamp="2026-02-20 08:26:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:26:42.667081947 +0000 UTC m=+1251.641576837" watchObservedRunningTime="2026-02-20 08:26:42.679075083 +0000 UTC m=+1251.653569933" Feb 20 08:26:50 crc kubenswrapper[4948]: I0220 08:26:50.133237 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:50 crc kubenswrapper[4948]: I0220 08:26:50.229745 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-vgk45"] Feb 20 08:26:50 crc kubenswrapper[4948]: I0220 08:26:50.231284 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" podUID="b4bae750-8bf9-4804-bab7-9c94d4be7bb6" containerName="dnsmasq-dns" containerID="cri-o://ede14cf7eee35beb81bccaf96c5abbd270f76ba98b65d67579bbb482b130f45e" gracePeriod=10 Feb 20 08:26:50 crc kubenswrapper[4948]: I0220 08:26:50.729353 4948 generic.go:334] "Generic (PLEG): container finished" podID="b4bae750-8bf9-4804-bab7-9c94d4be7bb6" containerID="ede14cf7eee35beb81bccaf96c5abbd270f76ba98b65d67579bbb482b130f45e" exitCode=0 Feb 20 08:26:50 crc kubenswrapper[4948]: I0220 08:26:50.729474 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" event={"ID":"b4bae750-8bf9-4804-bab7-9c94d4be7bb6","Type":"ContainerDied","Data":"ede14cf7eee35beb81bccaf96c5abbd270f76ba98b65d67579bbb482b130f45e"} Feb 20 08:26:50 crc kubenswrapper[4948]: I0220 08:26:50.729710 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" event={"ID":"b4bae750-8bf9-4804-bab7-9c94d4be7bb6","Type":"ContainerDied","Data":"a5fc6b8697916867d170463ef9a68553f0374e86750fe3d5b67599ef15a68b91"} Feb 20 08:26:50 crc kubenswrapper[4948]: I0220 08:26:50.729726 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a5fc6b8697916867d170463ef9a68553f0374e86750fe3d5b67599ef15a68b91" Feb 20 08:26:50 crc kubenswrapper[4948]: I0220 08:26:50.795573 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:50 crc kubenswrapper[4948]: I0220 08:26:50.980535 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-config\") pod \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " Feb 20 08:26:50 crc kubenswrapper[4948]: I0220 08:26:50.980608 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-ovsdbserver-nb\") pod \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " Feb 20 08:26:50 crc kubenswrapper[4948]: I0220 08:26:50.980643 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-dns-svc\") pod \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " Feb 20 08:26:50 crc kubenswrapper[4948]: I0220 08:26:50.980679 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-ovsdbserver-sb\") pod \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " Feb 20 08:26:50 crc kubenswrapper[4948]: I0220 08:26:50.980710 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-dns-swift-storage-0\") pod \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " Feb 20 08:26:50 crc kubenswrapper[4948]: I0220 08:26:50.980831 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4h8d\" (UniqueName: \"kubernetes.io/projected/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-kube-api-access-x4h8d\") pod \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " Feb 20 08:26:50 crc kubenswrapper[4948]: I0220 08:26:50.980873 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-openstack-edpm-ipam\") pod \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " Feb 20 08:26:51 crc kubenswrapper[4948]: I0220 08:26:51.010171 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-kube-api-access-x4h8d" (OuterVolumeSpecName: "kube-api-access-x4h8d") pod "b4bae750-8bf9-4804-bab7-9c94d4be7bb6" (UID: "b4bae750-8bf9-4804-bab7-9c94d4be7bb6"). InnerVolumeSpecName "kube-api-access-x4h8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:26:51 crc kubenswrapper[4948]: I0220 08:26:51.084034 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4h8d\" (UniqueName: \"kubernetes.io/projected/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-kube-api-access-x4h8d\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:51 crc kubenswrapper[4948]: I0220 08:26:51.104496 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b4bae750-8bf9-4804-bab7-9c94d4be7bb6" (UID: "b4bae750-8bf9-4804-bab7-9c94d4be7bb6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:26:51 crc kubenswrapper[4948]: I0220 08:26:51.110342 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b4bae750-8bf9-4804-bab7-9c94d4be7bb6" (UID: "b4bae750-8bf9-4804-bab7-9c94d4be7bb6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:26:51 crc kubenswrapper[4948]: I0220 08:26:51.116793 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "b4bae750-8bf9-4804-bab7-9c94d4be7bb6" (UID: "b4bae750-8bf9-4804-bab7-9c94d4be7bb6"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:26:51 crc kubenswrapper[4948]: I0220 08:26:51.126572 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b4bae750-8bf9-4804-bab7-9c94d4be7bb6" (UID: "b4bae750-8bf9-4804-bab7-9c94d4be7bb6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:26:51 crc kubenswrapper[4948]: I0220 08:26:51.142795 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b4bae750-8bf9-4804-bab7-9c94d4be7bb6" (UID: "b4bae750-8bf9-4804-bab7-9c94d4be7bb6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:26:51 crc kubenswrapper[4948]: I0220 08:26:51.147572 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-config" (OuterVolumeSpecName: "config") pod "b4bae750-8bf9-4804-bab7-9c94d4be7bb6" (UID: "b4bae750-8bf9-4804-bab7-9c94d4be7bb6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:26:51 crc kubenswrapper[4948]: I0220 08:26:51.185811 4948 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:51 crc kubenswrapper[4948]: I0220 08:26:51.185860 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:51 crc kubenswrapper[4948]: I0220 08:26:51.185869 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:51 crc kubenswrapper[4948]: I0220 08:26:51.185879 4948 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:51 crc kubenswrapper[4948]: I0220 08:26:51.185889 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:51 crc kubenswrapper[4948]: I0220 08:26:51.185897 4948 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:51 crc kubenswrapper[4948]: I0220 08:26:51.740366 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:51 crc kubenswrapper[4948]: I0220 08:26:51.804213 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-vgk45"] Feb 20 08:26:51 crc kubenswrapper[4948]: I0220 08:26:51.815682 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-vgk45"] Feb 20 08:26:53 crc kubenswrapper[4948]: I0220 08:26:53.737192 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4bae750-8bf9-4804-bab7-9c94d4be7bb6" path="/var/lib/kubelet/pods/b4bae750-8bf9-4804-bab7-9c94d4be7bb6/volumes" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.059885 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch"] Feb 20 08:27:03 crc kubenswrapper[4948]: E0220 08:27:03.060920 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1847849-cdf4-4ea6-a540-c01503a003b2" containerName="init" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.060935 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1847849-cdf4-4ea6-a540-c01503a003b2" containerName="init" Feb 20 08:27:03 crc kubenswrapper[4948]: E0220 08:27:03.060951 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4bae750-8bf9-4804-bab7-9c94d4be7bb6" containerName="dnsmasq-dns" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.060958 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4bae750-8bf9-4804-bab7-9c94d4be7bb6" containerName="dnsmasq-dns" Feb 20 08:27:03 crc kubenswrapper[4948]: E0220 08:27:03.060996 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4bae750-8bf9-4804-bab7-9c94d4be7bb6" containerName="init" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.061006 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4bae750-8bf9-4804-bab7-9c94d4be7bb6" containerName="init" Feb 20 08:27:03 crc kubenswrapper[4948]: E0220 08:27:03.061023 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1847849-cdf4-4ea6-a540-c01503a003b2" containerName="dnsmasq-dns" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.061030 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1847849-cdf4-4ea6-a540-c01503a003b2" containerName="dnsmasq-dns" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.061241 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4bae750-8bf9-4804-bab7-9c94d4be7bb6" containerName="dnsmasq-dns" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.061274 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1847849-cdf4-4ea6-a540-c01503a003b2" containerName="dnsmasq-dns" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.061942 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.065179 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.066921 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.069586 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.070244 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sgx49" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.094095 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch"] Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.160037 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch\" (UID: \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.160225 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch\" (UID: \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.160302 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch\" (UID: \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.160328 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2krxb\" (UniqueName: \"kubernetes.io/projected/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-kube-api-access-2krxb\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch\" (UID: \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.262339 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch\" (UID: \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.262395 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2krxb\" (UniqueName: \"kubernetes.io/projected/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-kube-api-access-2krxb\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch\" (UID: \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.262563 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch\" (UID: \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.262669 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch\" (UID: \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.268887 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch\" (UID: \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.269590 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch\" (UID: \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.271802 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch\" (UID: \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.281562 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2krxb\" (UniqueName: \"kubernetes.io/projected/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-kube-api-access-2krxb\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch\" (UID: \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.409593 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.896578 4948 generic.go:334] "Generic (PLEG): container finished" podID="037b329f-712f-4a67-984a-75affd2a57b7" containerID="9fcd3d9744d41147b3cfe90374517e1e56f49253e12cf9785df1944e69f8e3a2" exitCode=0 Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.896639 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"037b329f-712f-4a67-984a-75affd2a57b7","Type":"ContainerDied","Data":"9fcd3d9744d41147b3cfe90374517e1e56f49253e12cf9785df1944e69f8e3a2"} Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.903642 4948 generic.go:334] "Generic (PLEG): container finished" podID="f063bb62-10c3-46d9-a41d-56d2826059e0" containerID="878f9968270125c43cab5cdbc68ca06efa6e152bfb9c605c96475afc50f8ad38" exitCode=0 Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.903689 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f063bb62-10c3-46d9-a41d-56d2826059e0","Type":"ContainerDied","Data":"878f9968270125c43cab5cdbc68ca06efa6e152bfb9c605c96475afc50f8ad38"} Feb 20 08:27:04 crc kubenswrapper[4948]: I0220 08:27:04.025074 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch"] Feb 20 08:27:04 crc kubenswrapper[4948]: I0220 08:27:04.025764 4948 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 08:27:04 crc kubenswrapper[4948]: I0220 08:27:04.915849 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"037b329f-712f-4a67-984a-75affd2a57b7","Type":"ContainerStarted","Data":"9f638c6325085bb3f636e25b3784184a35ec0ac8c3fe43494b719b19a5a49e3a"} Feb 20 08:27:04 crc kubenswrapper[4948]: I0220 08:27:04.916413 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:27:04 crc kubenswrapper[4948]: I0220 08:27:04.917714 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" event={"ID":"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f","Type":"ContainerStarted","Data":"e390522c20f722c1095e8ce15bfbab5e5bebdf33d9f0c27bcc6e6d315e7becc5"} Feb 20 08:27:04 crc kubenswrapper[4948]: I0220 08:27:04.930006 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f063bb62-10c3-46d9-a41d-56d2826059e0","Type":"ContainerStarted","Data":"0366bfe9eabe6a009ea2dc265496f54894233021b297a92a54a5dc68da912163"} Feb 20 08:27:04 crc kubenswrapper[4948]: I0220 08:27:04.930381 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 20 08:27:04 crc kubenswrapper[4948]: I0220 08:27:04.940272 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.94025592 podStartE2EDuration="36.94025592s" podCreationTimestamp="2026-02-20 08:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:27:04.936865026 +0000 UTC m=+1273.911359846" watchObservedRunningTime="2026-02-20 08:27:04.94025592 +0000 UTC m=+1273.914750740" Feb 20 08:27:04 crc kubenswrapper[4948]: I0220 08:27:04.969255 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.969236705 podStartE2EDuration="37.969236705s" podCreationTimestamp="2026-02-20 08:26:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:27:04.966820825 +0000 UTC m=+1273.941315645" watchObservedRunningTime="2026-02-20 08:27:04.969236705 +0000 UTC m=+1273.943731525" Feb 20 08:27:08 crc kubenswrapper[4948]: I0220 08:27:08.025012 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:27:08 crc kubenswrapper[4948]: I0220 08:27:08.025376 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:27:14 crc kubenswrapper[4948]: I0220 08:27:14.028693 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" event={"ID":"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f","Type":"ContainerStarted","Data":"db62d9363c40e171b052da3433783d0ee6d1a1d0ff212dae6ec69ed654f36fb5"} Feb 20 08:27:18 crc kubenswrapper[4948]: I0220 08:27:18.132304 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 20 08:27:18 crc kubenswrapper[4948]: I0220 08:27:18.162518 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" podStartSLOduration=5.708976495 podStartE2EDuration="15.162500021s" podCreationTimestamp="2026-02-20 08:27:03 +0000 UTC" firstStartedPulling="2026-02-20 08:27:04.025499229 +0000 UTC m=+1272.999994059" lastFinishedPulling="2026-02-20 08:27:13.479022735 +0000 UTC m=+1282.453517585" observedRunningTime="2026-02-20 08:27:14.060010871 +0000 UTC m=+1283.034505731" watchObservedRunningTime="2026-02-20 08:27:18.162500021 +0000 UTC m=+1287.136994831" Feb 20 08:27:18 crc kubenswrapper[4948]: I0220 08:27:18.825134 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:27:24 crc kubenswrapper[4948]: I0220 08:27:24.186269 4948 generic.go:334] "Generic (PLEG): container finished" podID="90b6e7df-b3e1-4d05-a6b3-7fe47d70804f" containerID="db62d9363c40e171b052da3433783d0ee6d1a1d0ff212dae6ec69ed654f36fb5" exitCode=0 Feb 20 08:27:24 crc kubenswrapper[4948]: I0220 08:27:24.186368 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" event={"ID":"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f","Type":"ContainerDied","Data":"db62d9363c40e171b052da3433783d0ee6d1a1d0ff212dae6ec69ed654f36fb5"} Feb 20 08:27:25 crc kubenswrapper[4948]: I0220 08:27:25.674370 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" Feb 20 08:27:25 crc kubenswrapper[4948]: I0220 08:27:25.729481 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2krxb\" (UniqueName: \"kubernetes.io/projected/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-kube-api-access-2krxb\") pod \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\" (UID: \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\") " Feb 20 08:27:25 crc kubenswrapper[4948]: I0220 08:27:25.729680 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-repo-setup-combined-ca-bundle\") pod \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\" (UID: \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\") " Feb 20 08:27:25 crc kubenswrapper[4948]: I0220 08:27:25.729708 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-inventory\") pod \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\" (UID: \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\") " Feb 20 08:27:25 crc kubenswrapper[4948]: I0220 08:27:25.729750 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-ssh-key-openstack-edpm-ipam\") pod \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\" (UID: \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\") " Feb 20 08:27:25 crc kubenswrapper[4948]: I0220 08:27:25.735778 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "90b6e7df-b3e1-4d05-a6b3-7fe47d70804f" (UID: "90b6e7df-b3e1-4d05-a6b3-7fe47d70804f"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:27:25 crc kubenswrapper[4948]: I0220 08:27:25.740168 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-kube-api-access-2krxb" (OuterVolumeSpecName: "kube-api-access-2krxb") pod "90b6e7df-b3e1-4d05-a6b3-7fe47d70804f" (UID: "90b6e7df-b3e1-4d05-a6b3-7fe47d70804f"). InnerVolumeSpecName "kube-api-access-2krxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:27:25 crc kubenswrapper[4948]: I0220 08:27:25.762325 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "90b6e7df-b3e1-4d05-a6b3-7fe47d70804f" (UID: "90b6e7df-b3e1-4d05-a6b3-7fe47d70804f"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:27:25 crc kubenswrapper[4948]: I0220 08:27:25.769085 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-inventory" (OuterVolumeSpecName: "inventory") pod "90b6e7df-b3e1-4d05-a6b3-7fe47d70804f" (UID: "90b6e7df-b3e1-4d05-a6b3-7fe47d70804f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:27:25 crc kubenswrapper[4948]: I0220 08:27:25.833686 4948 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:27:25 crc kubenswrapper[4948]: I0220 08:27:25.833742 4948 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 08:27:25 crc kubenswrapper[4948]: I0220 08:27:25.833758 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:27:25 crc kubenswrapper[4948]: I0220 08:27:25.833773 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2krxb\" (UniqueName: \"kubernetes.io/projected/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-kube-api-access-2krxb\") on node \"crc\" DevicePath \"\"" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.211257 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" event={"ID":"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f","Type":"ContainerDied","Data":"e390522c20f722c1095e8ce15bfbab5e5bebdf33d9f0c27bcc6e6d315e7becc5"} Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.211617 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e390522c20f722c1095e8ce15bfbab5e5bebdf33d9f0c27bcc6e6d315e7becc5" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.211349 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.335418 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9"] Feb 20 08:27:26 crc kubenswrapper[4948]: E0220 08:27:26.335893 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90b6e7df-b3e1-4d05-a6b3-7fe47d70804f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.335914 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="90b6e7df-b3e1-4d05-a6b3-7fe47d70804f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.336137 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="90b6e7df-b3e1-4d05-a6b3-7fe47d70804f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.337036 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.339348 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sgx49" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.339865 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.340074 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.344595 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.349141 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9"] Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.450338 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdjs4\" (UniqueName: \"kubernetes.io/projected/c3e7cce4-5144-4190-8013-83428c66cde9-kube-api-access-sdjs4\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-f7ww9\" (UID: \"c3e7cce4-5144-4190-8013-83428c66cde9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.450693 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c3e7cce4-5144-4190-8013-83428c66cde9-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-f7ww9\" (UID: \"c3e7cce4-5144-4190-8013-83428c66cde9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.451056 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c3e7cce4-5144-4190-8013-83428c66cde9-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-f7ww9\" (UID: \"c3e7cce4-5144-4190-8013-83428c66cde9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.553058 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c3e7cce4-5144-4190-8013-83428c66cde9-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-f7ww9\" (UID: \"c3e7cce4-5144-4190-8013-83428c66cde9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.553473 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdjs4\" (UniqueName: \"kubernetes.io/projected/c3e7cce4-5144-4190-8013-83428c66cde9-kube-api-access-sdjs4\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-f7ww9\" (UID: \"c3e7cce4-5144-4190-8013-83428c66cde9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.553617 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c3e7cce4-5144-4190-8013-83428c66cde9-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-f7ww9\" (UID: \"c3e7cce4-5144-4190-8013-83428c66cde9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.561423 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c3e7cce4-5144-4190-8013-83428c66cde9-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-f7ww9\" (UID: \"c3e7cce4-5144-4190-8013-83428c66cde9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.562106 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c3e7cce4-5144-4190-8013-83428c66cde9-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-f7ww9\" (UID: \"c3e7cce4-5144-4190-8013-83428c66cde9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.570770 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdjs4\" (UniqueName: \"kubernetes.io/projected/c3e7cce4-5144-4190-8013-83428c66cde9-kube-api-access-sdjs4\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-f7ww9\" (UID: \"c3e7cce4-5144-4190-8013-83428c66cde9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.669739 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" Feb 20 08:27:27 crc kubenswrapper[4948]: I0220 08:27:27.290419 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9"] Feb 20 08:27:28 crc kubenswrapper[4948]: I0220 08:27:28.240919 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" event={"ID":"c3e7cce4-5144-4190-8013-83428c66cde9","Type":"ContainerStarted","Data":"3c99aa66568f6e79e574e82f764277e758470002a35a2b965f15c2be5d2ddbe6"} Feb 20 08:27:28 crc kubenswrapper[4948]: I0220 08:27:28.241409 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" event={"ID":"c3e7cce4-5144-4190-8013-83428c66cde9","Type":"ContainerStarted","Data":"92919d350e3d8fc6a7b429f20d4d8fc554e212055abe58d5ab447af9034b6250"} Feb 20 08:27:28 crc kubenswrapper[4948]: I0220 08:27:28.271717 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" podStartSLOduration=1.8622712940000001 podStartE2EDuration="2.271693857s" podCreationTimestamp="2026-02-20 08:27:26 +0000 UTC" firstStartedPulling="2026-02-20 08:27:27.299451976 +0000 UTC m=+1296.273946796" lastFinishedPulling="2026-02-20 08:27:27.708874539 +0000 UTC m=+1296.683369359" observedRunningTime="2026-02-20 08:27:28.264502699 +0000 UTC m=+1297.238997529" watchObservedRunningTime="2026-02-20 08:27:28.271693857 +0000 UTC m=+1297.246188687" Feb 20 08:27:31 crc kubenswrapper[4948]: I0220 08:27:31.290323 4948 generic.go:334] "Generic (PLEG): container finished" podID="c3e7cce4-5144-4190-8013-83428c66cde9" containerID="3c99aa66568f6e79e574e82f764277e758470002a35a2b965f15c2be5d2ddbe6" exitCode=0 Feb 20 08:27:31 crc kubenswrapper[4948]: I0220 08:27:31.290544 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" event={"ID":"c3e7cce4-5144-4190-8013-83428c66cde9","Type":"ContainerDied","Data":"3c99aa66568f6e79e574e82f764277e758470002a35a2b965f15c2be5d2ddbe6"} Feb 20 08:27:32 crc kubenswrapper[4948]: I0220 08:27:32.779582 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" Feb 20 08:27:32 crc kubenswrapper[4948]: I0220 08:27:32.891618 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c3e7cce4-5144-4190-8013-83428c66cde9-ssh-key-openstack-edpm-ipam\") pod \"c3e7cce4-5144-4190-8013-83428c66cde9\" (UID: \"c3e7cce4-5144-4190-8013-83428c66cde9\") " Feb 20 08:27:32 crc kubenswrapper[4948]: I0220 08:27:32.892130 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c3e7cce4-5144-4190-8013-83428c66cde9-inventory\") pod \"c3e7cce4-5144-4190-8013-83428c66cde9\" (UID: \"c3e7cce4-5144-4190-8013-83428c66cde9\") " Feb 20 08:27:32 crc kubenswrapper[4948]: I0220 08:27:32.892201 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdjs4\" (UniqueName: \"kubernetes.io/projected/c3e7cce4-5144-4190-8013-83428c66cde9-kube-api-access-sdjs4\") pod \"c3e7cce4-5144-4190-8013-83428c66cde9\" (UID: \"c3e7cce4-5144-4190-8013-83428c66cde9\") " Feb 20 08:27:32 crc kubenswrapper[4948]: I0220 08:27:32.901570 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3e7cce4-5144-4190-8013-83428c66cde9-kube-api-access-sdjs4" (OuterVolumeSpecName: "kube-api-access-sdjs4") pod "c3e7cce4-5144-4190-8013-83428c66cde9" (UID: "c3e7cce4-5144-4190-8013-83428c66cde9"). InnerVolumeSpecName "kube-api-access-sdjs4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:27:32 crc kubenswrapper[4948]: I0220 08:27:32.937045 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3e7cce4-5144-4190-8013-83428c66cde9-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "c3e7cce4-5144-4190-8013-83428c66cde9" (UID: "c3e7cce4-5144-4190-8013-83428c66cde9"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:27:32 crc kubenswrapper[4948]: I0220 08:27:32.944967 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3e7cce4-5144-4190-8013-83428c66cde9-inventory" (OuterVolumeSpecName: "inventory") pod "c3e7cce4-5144-4190-8013-83428c66cde9" (UID: "c3e7cce4-5144-4190-8013-83428c66cde9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:27:32 crc kubenswrapper[4948]: I0220 08:27:32.996035 4948 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c3e7cce4-5144-4190-8013-83428c66cde9-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 08:27:32 crc kubenswrapper[4948]: I0220 08:27:32.996097 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdjs4\" (UniqueName: \"kubernetes.io/projected/c3e7cce4-5144-4190-8013-83428c66cde9-kube-api-access-sdjs4\") on node \"crc\" DevicePath \"\"" Feb 20 08:27:32 crc kubenswrapper[4948]: I0220 08:27:32.996121 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c3e7cce4-5144-4190-8013-83428c66cde9-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.317787 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" event={"ID":"c3e7cce4-5144-4190-8013-83428c66cde9","Type":"ContainerDied","Data":"92919d350e3d8fc6a7b429f20d4d8fc554e212055abe58d5ab447af9034b6250"} Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.318227 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92919d350e3d8fc6a7b429f20d4d8fc554e212055abe58d5ab447af9034b6250" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.317859 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.432225 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g"] Feb 20 08:27:33 crc kubenswrapper[4948]: E0220 08:27:33.433230 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3e7cce4-5144-4190-8013-83428c66cde9" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.433261 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3e7cce4-5144-4190-8013-83428c66cde9" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.433597 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3e7cce4-5144-4190-8013-83428c66cde9" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.434689 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.441678 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.441869 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.441950 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.442049 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sgx49" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.448177 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g"] Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.508191 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4265g\" (UID: \"e9bed94f-dc85-433b-b2db-8da400959f54\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.508291 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4265g\" (UID: \"e9bed94f-dc85-433b-b2db-8da400959f54\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.508491 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljn9d\" (UniqueName: \"kubernetes.io/projected/e9bed94f-dc85-433b-b2db-8da400959f54-kube-api-access-ljn9d\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4265g\" (UID: \"e9bed94f-dc85-433b-b2db-8da400959f54\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.508865 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4265g\" (UID: \"e9bed94f-dc85-433b-b2db-8da400959f54\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.611155 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4265g\" (UID: \"e9bed94f-dc85-433b-b2db-8da400959f54\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.611270 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljn9d\" (UniqueName: \"kubernetes.io/projected/e9bed94f-dc85-433b-b2db-8da400959f54-kube-api-access-ljn9d\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4265g\" (UID: \"e9bed94f-dc85-433b-b2db-8da400959f54\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.611414 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4265g\" (UID: \"e9bed94f-dc85-433b-b2db-8da400959f54\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.611554 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4265g\" (UID: \"e9bed94f-dc85-433b-b2db-8da400959f54\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.616812 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4265g\" (UID: \"e9bed94f-dc85-433b-b2db-8da400959f54\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.623451 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4265g\" (UID: \"e9bed94f-dc85-433b-b2db-8da400959f54\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.623932 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4265g\" (UID: \"e9bed94f-dc85-433b-b2db-8da400959f54\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.634935 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljn9d\" (UniqueName: \"kubernetes.io/projected/e9bed94f-dc85-433b-b2db-8da400959f54-kube-api-access-ljn9d\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4265g\" (UID: \"e9bed94f-dc85-433b-b2db-8da400959f54\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.767446 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" Feb 20 08:27:34 crc kubenswrapper[4948]: I0220 08:27:34.342237 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g"] Feb 20 08:27:34 crc kubenswrapper[4948]: W0220 08:27:34.347437 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9bed94f_dc85_433b_b2db_8da400959f54.slice/crio-569d95f896c3d860d075c4d7112372cdd51f62d88d362789225ed9ac1821cb97 WatchSource:0}: Error finding container 569d95f896c3d860d075c4d7112372cdd51f62d88d362789225ed9ac1821cb97: Status 404 returned error can't find the container with id 569d95f896c3d860d075c4d7112372cdd51f62d88d362789225ed9ac1821cb97 Feb 20 08:27:35 crc kubenswrapper[4948]: I0220 08:27:35.351665 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" event={"ID":"e9bed94f-dc85-433b-b2db-8da400959f54","Type":"ContainerStarted","Data":"966bce30bb61203264725c1e50c5639d590639efbe4b4a138802094aa4f1d142"} Feb 20 08:27:35 crc kubenswrapper[4948]: I0220 08:27:35.351944 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" event={"ID":"e9bed94f-dc85-433b-b2db-8da400959f54","Type":"ContainerStarted","Data":"569d95f896c3d860d075c4d7112372cdd51f62d88d362789225ed9ac1821cb97"} Feb 20 08:27:38 crc kubenswrapper[4948]: I0220 08:27:38.025063 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:27:38 crc kubenswrapper[4948]: I0220 08:27:38.025778 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:27:38 crc kubenswrapper[4948]: I0220 08:27:38.025846 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:27:38 crc kubenswrapper[4948]: I0220 08:27:38.026727 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d9efd6716995fcdcebdcf01e76fc7927735e34b1da4918444f90e59bd0f6d815"} pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 08:27:38 crc kubenswrapper[4948]: I0220 08:27:38.026823 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" containerID="cri-o://d9efd6716995fcdcebdcf01e76fc7927735e34b1da4918444f90e59bd0f6d815" gracePeriod=600 Feb 20 08:27:38 crc kubenswrapper[4948]: I0220 08:27:38.390667 4948 generic.go:334] "Generic (PLEG): container finished" podID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerID="d9efd6716995fcdcebdcf01e76fc7927735e34b1da4918444f90e59bd0f6d815" exitCode=0 Feb 20 08:27:38 crc kubenswrapper[4948]: I0220 08:27:38.390758 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerDied","Data":"d9efd6716995fcdcebdcf01e76fc7927735e34b1da4918444f90e59bd0f6d815"} Feb 20 08:27:38 crc kubenswrapper[4948]: I0220 08:27:38.391482 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43"} Feb 20 08:27:38 crc kubenswrapper[4948]: I0220 08:27:38.391546 4948 scope.go:117] "RemoveContainer" containerID="8f0b8a806330c630468951158113ee0ecaad5dbddb81b09c063752a17f21282c" Feb 20 08:27:38 crc kubenswrapper[4948]: I0220 08:27:38.424231 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" podStartSLOduration=5.00794339 podStartE2EDuration="5.424211271s" podCreationTimestamp="2026-02-20 08:27:33 +0000 UTC" firstStartedPulling="2026-02-20 08:27:34.350785309 +0000 UTC m=+1303.325280129" lastFinishedPulling="2026-02-20 08:27:34.76705318 +0000 UTC m=+1303.741548010" observedRunningTime="2026-02-20 08:27:35.376701823 +0000 UTC m=+1304.351196703" watchObservedRunningTime="2026-02-20 08:27:38.424211271 +0000 UTC m=+1307.398706091" Feb 20 08:27:57 crc kubenswrapper[4948]: I0220 08:27:57.961889 4948 scope.go:117] "RemoveContainer" containerID="562bf68215847fe83f99073d009b1ee89904610f3d702b2ab282baecbca78fb2" Feb 20 08:27:58 crc kubenswrapper[4948]: I0220 08:27:58.002130 4948 scope.go:117] "RemoveContainer" containerID="d4c99094574230bbe4d4f5bd6cf1019d664fb9f16355c0c0e8d85cd644e73d32" Feb 20 08:27:58 crc kubenswrapper[4948]: I0220 08:27:58.047844 4948 scope.go:117] "RemoveContainer" containerID="48592a00a61a9ce774a176445c59d2e205743cbfe0a6d5b932c49a8ff31d312a" Feb 20 08:27:58 crc kubenswrapper[4948]: I0220 08:27:58.084311 4948 scope.go:117] "RemoveContainer" containerID="c4af4c0a8310dde22c0962922d9e1a2de48ad4ce8cecfbf27d7c0a78b93a2059" Feb 20 08:27:58 crc kubenswrapper[4948]: I0220 08:27:58.125330 4948 scope.go:117] "RemoveContainer" containerID="5dfd385e001fae733c595da928ba9b46f6e3cc2e4d358e37840e650d21ce855a" Feb 20 08:28:58 crc kubenswrapper[4948]: I0220 08:28:58.274456 4948 scope.go:117] "RemoveContainer" containerID="656e84496ca041e8a454b07ed67e165b130af1030016564ae4519b8e2e6f11e5" Feb 20 08:28:58 crc kubenswrapper[4948]: I0220 08:28:58.332901 4948 scope.go:117] "RemoveContainer" containerID="577865ca69f0d8457d617948e5cac2be0a124e4620edc60df1395130de8a2334" Feb 20 08:28:58 crc kubenswrapper[4948]: I0220 08:28:58.355680 4948 scope.go:117] "RemoveContainer" containerID="a1b60a312d88b84db5225c1aff5290457499b2775b6fffec7ca3246896b28523" Feb 20 08:28:58 crc kubenswrapper[4948]: I0220 08:28:58.593685 4948 scope.go:117] "RemoveContainer" containerID="fa275de1b67d8651c437fb815c4f8470dc8d77cd9885aaddee1a7419f33fd25c" Feb 20 08:28:58 crc kubenswrapper[4948]: I0220 08:28:58.656936 4948 scope.go:117] "RemoveContainer" containerID="344f41da66c91718be93d872d218e57c352e04b55c97d3dc69d71b942dbe58d1" Feb 20 08:28:58 crc kubenswrapper[4948]: I0220 08:28:58.682561 4948 scope.go:117] "RemoveContainer" containerID="a66ddb00a7f57dc492a3a8a03a3bce06f45c65381b1e5db4f923e10fec5267a0" Feb 20 08:28:58 crc kubenswrapper[4948]: I0220 08:28:58.714102 4948 scope.go:117] "RemoveContainer" containerID="e38f809977ba5076c8af037b7e38d52d33545f893ce90480e346caeaaef81dee" Feb 20 08:29:38 crc kubenswrapper[4948]: I0220 08:29:38.025596 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:29:38 crc kubenswrapper[4948]: I0220 08:29:38.027238 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:30:00 crc kubenswrapper[4948]: I0220 08:30:00.151947 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm"] Feb 20 08:30:00 crc kubenswrapper[4948]: I0220 08:30:00.153870 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm" Feb 20 08:30:00 crc kubenswrapper[4948]: I0220 08:30:00.155720 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 20 08:30:00 crc kubenswrapper[4948]: I0220 08:30:00.165044 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 20 08:30:00 crc kubenswrapper[4948]: I0220 08:30:00.178993 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm"] Feb 20 08:30:00 crc kubenswrapper[4948]: I0220 08:30:00.310425 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpkkr\" (UniqueName: \"kubernetes.io/projected/2aaa3d0b-4371-404a-8d94-9a7158fff416-kube-api-access-hpkkr\") pod \"collect-profiles-29526270-98wxm\" (UID: \"2aaa3d0b-4371-404a-8d94-9a7158fff416\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm" Feb 20 08:30:00 crc kubenswrapper[4948]: I0220 08:30:00.310516 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2aaa3d0b-4371-404a-8d94-9a7158fff416-secret-volume\") pod \"collect-profiles-29526270-98wxm\" (UID: \"2aaa3d0b-4371-404a-8d94-9a7158fff416\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm" Feb 20 08:30:00 crc kubenswrapper[4948]: I0220 08:30:00.310597 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2aaa3d0b-4371-404a-8d94-9a7158fff416-config-volume\") pod \"collect-profiles-29526270-98wxm\" (UID: \"2aaa3d0b-4371-404a-8d94-9a7158fff416\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm" Feb 20 08:30:00 crc kubenswrapper[4948]: I0220 08:30:00.412799 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2aaa3d0b-4371-404a-8d94-9a7158fff416-config-volume\") pod \"collect-profiles-29526270-98wxm\" (UID: \"2aaa3d0b-4371-404a-8d94-9a7158fff416\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm" Feb 20 08:30:00 crc kubenswrapper[4948]: I0220 08:30:00.413112 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpkkr\" (UniqueName: \"kubernetes.io/projected/2aaa3d0b-4371-404a-8d94-9a7158fff416-kube-api-access-hpkkr\") pod \"collect-profiles-29526270-98wxm\" (UID: \"2aaa3d0b-4371-404a-8d94-9a7158fff416\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm" Feb 20 08:30:00 crc kubenswrapper[4948]: I0220 08:30:00.413156 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2aaa3d0b-4371-404a-8d94-9a7158fff416-secret-volume\") pod \"collect-profiles-29526270-98wxm\" (UID: \"2aaa3d0b-4371-404a-8d94-9a7158fff416\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm" Feb 20 08:30:00 crc kubenswrapper[4948]: I0220 08:30:00.414374 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2aaa3d0b-4371-404a-8d94-9a7158fff416-config-volume\") pod \"collect-profiles-29526270-98wxm\" (UID: \"2aaa3d0b-4371-404a-8d94-9a7158fff416\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm" Feb 20 08:30:00 crc kubenswrapper[4948]: I0220 08:30:00.419120 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2aaa3d0b-4371-404a-8d94-9a7158fff416-secret-volume\") pod \"collect-profiles-29526270-98wxm\" (UID: \"2aaa3d0b-4371-404a-8d94-9a7158fff416\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm" Feb 20 08:30:00 crc kubenswrapper[4948]: I0220 08:30:00.433017 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpkkr\" (UniqueName: \"kubernetes.io/projected/2aaa3d0b-4371-404a-8d94-9a7158fff416-kube-api-access-hpkkr\") pod \"collect-profiles-29526270-98wxm\" (UID: \"2aaa3d0b-4371-404a-8d94-9a7158fff416\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm" Feb 20 08:30:00 crc kubenswrapper[4948]: I0220 08:30:00.473949 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm" Feb 20 08:30:00 crc kubenswrapper[4948]: I0220 08:30:00.986612 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm"] Feb 20 08:30:01 crc kubenswrapper[4948]: I0220 08:30:01.138683 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm" event={"ID":"2aaa3d0b-4371-404a-8d94-9a7158fff416","Type":"ContainerStarted","Data":"34ee96251b37f911907f4a06d490080e0cb865adf32db4a0b0cd8c0d113ad82a"} Feb 20 08:30:02 crc kubenswrapper[4948]: I0220 08:30:02.153592 4948 generic.go:334] "Generic (PLEG): container finished" podID="2aaa3d0b-4371-404a-8d94-9a7158fff416" containerID="21765550ece8828e59342442ce6d5be7d08bd1dbcd30f8abbcc286c1bfd5f713" exitCode=0 Feb 20 08:30:02 crc kubenswrapper[4948]: I0220 08:30:02.153722 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm" event={"ID":"2aaa3d0b-4371-404a-8d94-9a7158fff416","Type":"ContainerDied","Data":"21765550ece8828e59342442ce6d5be7d08bd1dbcd30f8abbcc286c1bfd5f713"} Feb 20 08:30:03 crc kubenswrapper[4948]: I0220 08:30:03.503152 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm" Feb 20 08:30:03 crc kubenswrapper[4948]: I0220 08:30:03.675943 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpkkr\" (UniqueName: \"kubernetes.io/projected/2aaa3d0b-4371-404a-8d94-9a7158fff416-kube-api-access-hpkkr\") pod \"2aaa3d0b-4371-404a-8d94-9a7158fff416\" (UID: \"2aaa3d0b-4371-404a-8d94-9a7158fff416\") " Feb 20 08:30:03 crc kubenswrapper[4948]: I0220 08:30:03.676086 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2aaa3d0b-4371-404a-8d94-9a7158fff416-secret-volume\") pod \"2aaa3d0b-4371-404a-8d94-9a7158fff416\" (UID: \"2aaa3d0b-4371-404a-8d94-9a7158fff416\") " Feb 20 08:30:03 crc kubenswrapper[4948]: I0220 08:30:03.676214 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2aaa3d0b-4371-404a-8d94-9a7158fff416-config-volume\") pod \"2aaa3d0b-4371-404a-8d94-9a7158fff416\" (UID: \"2aaa3d0b-4371-404a-8d94-9a7158fff416\") " Feb 20 08:30:03 crc kubenswrapper[4948]: I0220 08:30:03.677122 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2aaa3d0b-4371-404a-8d94-9a7158fff416-config-volume" (OuterVolumeSpecName: "config-volume") pod "2aaa3d0b-4371-404a-8d94-9a7158fff416" (UID: "2aaa3d0b-4371-404a-8d94-9a7158fff416"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:30:03 crc kubenswrapper[4948]: I0220 08:30:03.685890 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2aaa3d0b-4371-404a-8d94-9a7158fff416-kube-api-access-hpkkr" (OuterVolumeSpecName: "kube-api-access-hpkkr") pod "2aaa3d0b-4371-404a-8d94-9a7158fff416" (UID: "2aaa3d0b-4371-404a-8d94-9a7158fff416"). InnerVolumeSpecName "kube-api-access-hpkkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:30:03 crc kubenswrapper[4948]: I0220 08:30:03.686127 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aaa3d0b-4371-404a-8d94-9a7158fff416-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2aaa3d0b-4371-404a-8d94-9a7158fff416" (UID: "2aaa3d0b-4371-404a-8d94-9a7158fff416"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:30:03 crc kubenswrapper[4948]: I0220 08:30:03.778825 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpkkr\" (UniqueName: \"kubernetes.io/projected/2aaa3d0b-4371-404a-8d94-9a7158fff416-kube-api-access-hpkkr\") on node \"crc\" DevicePath \"\"" Feb 20 08:30:03 crc kubenswrapper[4948]: I0220 08:30:03.778877 4948 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2aaa3d0b-4371-404a-8d94-9a7158fff416-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 20 08:30:03 crc kubenswrapper[4948]: I0220 08:30:03.778898 4948 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2aaa3d0b-4371-404a-8d94-9a7158fff416-config-volume\") on node \"crc\" DevicePath \"\"" Feb 20 08:30:04 crc kubenswrapper[4948]: I0220 08:30:04.181960 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm" event={"ID":"2aaa3d0b-4371-404a-8d94-9a7158fff416","Type":"ContainerDied","Data":"34ee96251b37f911907f4a06d490080e0cb865adf32db4a0b0cd8c0d113ad82a"} Feb 20 08:30:04 crc kubenswrapper[4948]: I0220 08:30:04.182215 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34ee96251b37f911907f4a06d490080e0cb865adf32db4a0b0cd8c0d113ad82a" Feb 20 08:30:04 crc kubenswrapper[4948]: I0220 08:30:04.182560 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm" Feb 20 08:30:08 crc kubenswrapper[4948]: I0220 08:30:08.024746 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:30:08 crc kubenswrapper[4948]: I0220 08:30:08.025424 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:30:26 crc kubenswrapper[4948]: I0220 08:30:26.404256 4948 generic.go:334] "Generic (PLEG): container finished" podID="e9bed94f-dc85-433b-b2db-8da400959f54" containerID="966bce30bb61203264725c1e50c5639d590639efbe4b4a138802094aa4f1d142" exitCode=0 Feb 20 08:30:26 crc kubenswrapper[4948]: I0220 08:30:26.404372 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" event={"ID":"e9bed94f-dc85-433b-b2db-8da400959f54","Type":"ContainerDied","Data":"966bce30bb61203264725c1e50c5639d590639efbe4b4a138802094aa4f1d142"} Feb 20 08:30:27 crc kubenswrapper[4948]: I0220 08:30:27.913182 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.086391 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-inventory\") pod \"e9bed94f-dc85-433b-b2db-8da400959f54\" (UID: \"e9bed94f-dc85-433b-b2db-8da400959f54\") " Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.086465 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-ssh-key-openstack-edpm-ipam\") pod \"e9bed94f-dc85-433b-b2db-8da400959f54\" (UID: \"e9bed94f-dc85-433b-b2db-8da400959f54\") " Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.086547 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-bootstrap-combined-ca-bundle\") pod \"e9bed94f-dc85-433b-b2db-8da400959f54\" (UID: \"e9bed94f-dc85-433b-b2db-8da400959f54\") " Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.086577 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljn9d\" (UniqueName: \"kubernetes.io/projected/e9bed94f-dc85-433b-b2db-8da400959f54-kube-api-access-ljn9d\") pod \"e9bed94f-dc85-433b-b2db-8da400959f54\" (UID: \"e9bed94f-dc85-433b-b2db-8da400959f54\") " Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.092169 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "e9bed94f-dc85-433b-b2db-8da400959f54" (UID: "e9bed94f-dc85-433b-b2db-8da400959f54"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.097966 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9bed94f-dc85-433b-b2db-8da400959f54-kube-api-access-ljn9d" (OuterVolumeSpecName: "kube-api-access-ljn9d") pod "e9bed94f-dc85-433b-b2db-8da400959f54" (UID: "e9bed94f-dc85-433b-b2db-8da400959f54"). InnerVolumeSpecName "kube-api-access-ljn9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.113590 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-inventory" (OuterVolumeSpecName: "inventory") pod "e9bed94f-dc85-433b-b2db-8da400959f54" (UID: "e9bed94f-dc85-433b-b2db-8da400959f54"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.117552 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "e9bed94f-dc85-433b-b2db-8da400959f54" (UID: "e9bed94f-dc85-433b-b2db-8da400959f54"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.188504 4948 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.188549 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.188560 4948 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.188569 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljn9d\" (UniqueName: \"kubernetes.io/projected/e9bed94f-dc85-433b-b2db-8da400959f54-kube-api-access-ljn9d\") on node \"crc\" DevicePath \"\"" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.436641 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" event={"ID":"e9bed94f-dc85-433b-b2db-8da400959f54","Type":"ContainerDied","Data":"569d95f896c3d860d075c4d7112372cdd51f62d88d362789225ed9ac1821cb97"} Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.436935 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="569d95f896c3d860d075c4d7112372cdd51f62d88d362789225ed9ac1821cb97" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.436718 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.542147 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll"] Feb 20 08:30:28 crc kubenswrapper[4948]: E0220 08:30:28.542558 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aaa3d0b-4371-404a-8d94-9a7158fff416" containerName="collect-profiles" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.542574 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aaa3d0b-4371-404a-8d94-9a7158fff416" containerName="collect-profiles" Feb 20 08:30:28 crc kubenswrapper[4948]: E0220 08:30:28.542614 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9bed94f-dc85-433b-b2db-8da400959f54" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.542623 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9bed94f-dc85-433b-b2db-8da400959f54" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.542796 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9bed94f-dc85-433b-b2db-8da400959f54" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.542816 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="2aaa3d0b-4371-404a-8d94-9a7158fff416" containerName="collect-profiles" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.543707 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.547949 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.548473 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.548714 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.548920 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sgx49" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.567651 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll"] Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.697845 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgbgc\" (UniqueName: \"kubernetes.io/projected/5864a68d-650e-4bcf-b705-619c0f27445b-kube-api-access-lgbgc\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-mhkll\" (UID: \"5864a68d-650e-4bcf-b705-619c0f27445b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.697965 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5864a68d-650e-4bcf-b705-619c0f27445b-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-mhkll\" (UID: \"5864a68d-650e-4bcf-b705-619c0f27445b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.698167 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5864a68d-650e-4bcf-b705-619c0f27445b-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-mhkll\" (UID: \"5864a68d-650e-4bcf-b705-619c0f27445b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.800434 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5864a68d-650e-4bcf-b705-619c0f27445b-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-mhkll\" (UID: \"5864a68d-650e-4bcf-b705-619c0f27445b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.800555 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5864a68d-650e-4bcf-b705-619c0f27445b-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-mhkll\" (UID: \"5864a68d-650e-4bcf-b705-619c0f27445b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.800831 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgbgc\" (UniqueName: \"kubernetes.io/projected/5864a68d-650e-4bcf-b705-619c0f27445b-kube-api-access-lgbgc\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-mhkll\" (UID: \"5864a68d-650e-4bcf-b705-619c0f27445b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.807219 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5864a68d-650e-4bcf-b705-619c0f27445b-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-mhkll\" (UID: \"5864a68d-650e-4bcf-b705-619c0f27445b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.808759 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5864a68d-650e-4bcf-b705-619c0f27445b-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-mhkll\" (UID: \"5864a68d-650e-4bcf-b705-619c0f27445b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.831187 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgbgc\" (UniqueName: \"kubernetes.io/projected/5864a68d-650e-4bcf-b705-619c0f27445b-kube-api-access-lgbgc\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-mhkll\" (UID: \"5864a68d-650e-4bcf-b705-619c0f27445b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.872905 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" Feb 20 08:30:29 crc kubenswrapper[4948]: I0220 08:30:29.473436 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll"] Feb 20 08:30:29 crc kubenswrapper[4948]: W0220 08:30:29.487390 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5864a68d_650e_4bcf_b705_619c0f27445b.slice/crio-3ff23193cbb1aafb8b5d373010ecf579e6350e550af86a1ff2fe999d6f3221b3 WatchSource:0}: Error finding container 3ff23193cbb1aafb8b5d373010ecf579e6350e550af86a1ff2fe999d6f3221b3: Status 404 returned error can't find the container with id 3ff23193cbb1aafb8b5d373010ecf579e6350e550af86a1ff2fe999d6f3221b3 Feb 20 08:30:30 crc kubenswrapper[4948]: I0220 08:30:30.457290 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" event={"ID":"5864a68d-650e-4bcf-b705-619c0f27445b","Type":"ContainerStarted","Data":"bb15461fc1a5607b927c4774ccf8aa52aa2b1f66a0749cf932076cda97280dfb"} Feb 20 08:30:30 crc kubenswrapper[4948]: I0220 08:30:30.457863 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" event={"ID":"5864a68d-650e-4bcf-b705-619c0f27445b","Type":"ContainerStarted","Data":"3ff23193cbb1aafb8b5d373010ecf579e6350e550af86a1ff2fe999d6f3221b3"} Feb 20 08:30:30 crc kubenswrapper[4948]: I0220 08:30:30.480546 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" podStartSLOduration=1.968606227 podStartE2EDuration="2.480524084s" podCreationTimestamp="2026-02-20 08:30:28 +0000 UTC" firstStartedPulling="2026-02-20 08:30:29.490123109 +0000 UTC m=+1478.464617929" lastFinishedPulling="2026-02-20 08:30:30.002040966 +0000 UTC m=+1478.976535786" observedRunningTime="2026-02-20 08:30:30.470715932 +0000 UTC m=+1479.445210752" watchObservedRunningTime="2026-02-20 08:30:30.480524084 +0000 UTC m=+1479.455018904" Feb 20 08:30:38 crc kubenswrapper[4948]: I0220 08:30:38.025710 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:30:38 crc kubenswrapper[4948]: I0220 08:30:38.026557 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:30:38 crc kubenswrapper[4948]: I0220 08:30:38.026681 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:30:38 crc kubenswrapper[4948]: I0220 08:30:38.028033 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43"} pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 08:30:38 crc kubenswrapper[4948]: I0220 08:30:38.028403 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" containerID="cri-o://abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" gracePeriod=600 Feb 20 08:30:38 crc kubenswrapper[4948]: E0220 08:30:38.154015 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:30:38 crc kubenswrapper[4948]: I0220 08:30:38.540311 4948 generic.go:334] "Generic (PLEG): container finished" podID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" exitCode=0 Feb 20 08:30:38 crc kubenswrapper[4948]: I0220 08:30:38.540371 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerDied","Data":"abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43"} Feb 20 08:30:38 crc kubenswrapper[4948]: I0220 08:30:38.540417 4948 scope.go:117] "RemoveContainer" containerID="d9efd6716995fcdcebdcf01e76fc7927735e34b1da4918444f90e59bd0f6d815" Feb 20 08:30:38 crc kubenswrapper[4948]: I0220 08:30:38.541267 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:30:38 crc kubenswrapper[4948]: E0220 08:30:38.541724 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:30:52 crc kubenswrapper[4948]: I0220 08:30:52.723197 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:30:52 crc kubenswrapper[4948]: E0220 08:30:52.725563 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:31:06 crc kubenswrapper[4948]: I0220 08:31:06.722951 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:31:06 crc kubenswrapper[4948]: E0220 08:31:06.723915 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:31:21 crc kubenswrapper[4948]: I0220 08:31:21.730231 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:31:21 crc kubenswrapper[4948]: E0220 08:31:21.731040 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:31:34 crc kubenswrapper[4948]: I0220 08:31:34.883503 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bdphs"] Feb 20 08:31:34 crc kubenswrapper[4948]: I0220 08:31:34.886606 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bdphs" Feb 20 08:31:34 crc kubenswrapper[4948]: I0220 08:31:34.901490 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bdphs"] Feb 20 08:31:34 crc kubenswrapper[4948]: I0220 08:31:34.998346 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/060989bf-ef06-4060-bc25-915180f42e5b-catalog-content\") pod \"redhat-operators-bdphs\" (UID: \"060989bf-ef06-4060-bc25-915180f42e5b\") " pod="openshift-marketplace/redhat-operators-bdphs" Feb 20 08:31:34 crc kubenswrapper[4948]: I0220 08:31:34.998414 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4l4qt\" (UniqueName: \"kubernetes.io/projected/060989bf-ef06-4060-bc25-915180f42e5b-kube-api-access-4l4qt\") pod \"redhat-operators-bdphs\" (UID: \"060989bf-ef06-4060-bc25-915180f42e5b\") " pod="openshift-marketplace/redhat-operators-bdphs" Feb 20 08:31:34 crc kubenswrapper[4948]: I0220 08:31:34.998494 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/060989bf-ef06-4060-bc25-915180f42e5b-utilities\") pod \"redhat-operators-bdphs\" (UID: \"060989bf-ef06-4060-bc25-915180f42e5b\") " pod="openshift-marketplace/redhat-operators-bdphs" Feb 20 08:31:35 crc kubenswrapper[4948]: I0220 08:31:35.100545 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/060989bf-ef06-4060-bc25-915180f42e5b-catalog-content\") pod \"redhat-operators-bdphs\" (UID: \"060989bf-ef06-4060-bc25-915180f42e5b\") " pod="openshift-marketplace/redhat-operators-bdphs" Feb 20 08:31:35 crc kubenswrapper[4948]: I0220 08:31:35.100697 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4l4qt\" (UniqueName: \"kubernetes.io/projected/060989bf-ef06-4060-bc25-915180f42e5b-kube-api-access-4l4qt\") pod \"redhat-operators-bdphs\" (UID: \"060989bf-ef06-4060-bc25-915180f42e5b\") " pod="openshift-marketplace/redhat-operators-bdphs" Feb 20 08:31:35 crc kubenswrapper[4948]: I0220 08:31:35.100933 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/060989bf-ef06-4060-bc25-915180f42e5b-utilities\") pod \"redhat-operators-bdphs\" (UID: \"060989bf-ef06-4060-bc25-915180f42e5b\") " pod="openshift-marketplace/redhat-operators-bdphs" Feb 20 08:31:35 crc kubenswrapper[4948]: I0220 08:31:35.101065 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/060989bf-ef06-4060-bc25-915180f42e5b-catalog-content\") pod \"redhat-operators-bdphs\" (UID: \"060989bf-ef06-4060-bc25-915180f42e5b\") " pod="openshift-marketplace/redhat-operators-bdphs" Feb 20 08:31:35 crc kubenswrapper[4948]: I0220 08:31:35.101584 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/060989bf-ef06-4060-bc25-915180f42e5b-utilities\") pod \"redhat-operators-bdphs\" (UID: \"060989bf-ef06-4060-bc25-915180f42e5b\") " pod="openshift-marketplace/redhat-operators-bdphs" Feb 20 08:31:35 crc kubenswrapper[4948]: I0220 08:31:35.120515 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4l4qt\" (UniqueName: \"kubernetes.io/projected/060989bf-ef06-4060-bc25-915180f42e5b-kube-api-access-4l4qt\") pod \"redhat-operators-bdphs\" (UID: \"060989bf-ef06-4060-bc25-915180f42e5b\") " pod="openshift-marketplace/redhat-operators-bdphs" Feb 20 08:31:35 crc kubenswrapper[4948]: I0220 08:31:35.256669 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bdphs" Feb 20 08:31:35 crc kubenswrapper[4948]: I0220 08:31:35.715275 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bdphs"] Feb 20 08:31:36 crc kubenswrapper[4948]: I0220 08:31:36.050405 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-bv65d"] Feb 20 08:31:36 crc kubenswrapper[4948]: I0220 08:31:36.055779 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-eb36-account-create-update-554g5"] Feb 20 08:31:36 crc kubenswrapper[4948]: I0220 08:31:36.064749 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-6d58-account-create-update-bbx44"] Feb 20 08:31:36 crc kubenswrapper[4948]: I0220 08:31:36.076611 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-gm6g6"] Feb 20 08:31:36 crc kubenswrapper[4948]: I0220 08:31:36.087538 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-bv65d"] Feb 20 08:31:36 crc kubenswrapper[4948]: I0220 08:31:36.095016 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-eb36-account-create-update-554g5"] Feb 20 08:31:36 crc kubenswrapper[4948]: I0220 08:31:36.120633 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-6d58-account-create-update-bbx44"] Feb 20 08:31:36 crc kubenswrapper[4948]: I0220 08:31:36.138198 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-gm6g6"] Feb 20 08:31:36 crc kubenswrapper[4948]: I0220 08:31:36.145676 4948 generic.go:334] "Generic (PLEG): container finished" podID="060989bf-ef06-4060-bc25-915180f42e5b" containerID="cafef04c966377e665f78102a5914bb6441b7f707970965b08dc70cfb2cf82a1" exitCode=0 Feb 20 08:31:36 crc kubenswrapper[4948]: I0220 08:31:36.145738 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdphs" event={"ID":"060989bf-ef06-4060-bc25-915180f42e5b","Type":"ContainerDied","Data":"cafef04c966377e665f78102a5914bb6441b7f707970965b08dc70cfb2cf82a1"} Feb 20 08:31:36 crc kubenswrapper[4948]: I0220 08:31:36.145764 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdphs" event={"ID":"060989bf-ef06-4060-bc25-915180f42e5b","Type":"ContainerStarted","Data":"53735cceaab8db6cbf509cf3d7a338fd07617ca1b94b6981218603729cf70256"} Feb 20 08:31:36 crc kubenswrapper[4948]: I0220 08:31:36.723539 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:31:36 crc kubenswrapper[4948]: E0220 08:31:36.723945 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:31:37 crc kubenswrapper[4948]: I0220 08:31:37.156534 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdphs" event={"ID":"060989bf-ef06-4060-bc25-915180f42e5b","Type":"ContainerStarted","Data":"7cfda758c2e1d88101427b25367b79d9870e89cc2808d38e69b1e5dfa5780966"} Feb 20 08:31:37 crc kubenswrapper[4948]: I0220 08:31:37.736221 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52238644-5964-427e-8b77-9ca228d031db" path="/var/lib/kubelet/pods/52238644-5964-427e-8b77-9ca228d031db/volumes" Feb 20 08:31:37 crc kubenswrapper[4948]: I0220 08:31:37.738563 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5bf7a21-af5d-4334-bc98-bd0f9eb9035a" path="/var/lib/kubelet/pods/b5bf7a21-af5d-4334-bc98-bd0f9eb9035a/volumes" Feb 20 08:31:37 crc kubenswrapper[4948]: I0220 08:31:37.739957 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb" path="/var/lib/kubelet/pods/e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb/volumes" Feb 20 08:31:37 crc kubenswrapper[4948]: I0220 08:31:37.741223 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f917fa7f-51f9-4c8f-bee2-11529a6d44a6" path="/var/lib/kubelet/pods/f917fa7f-51f9-4c8f-bee2-11529a6d44a6/volumes" Feb 20 08:31:38 crc kubenswrapper[4948]: I0220 08:31:38.040791 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-fs9rr"] Feb 20 08:31:38 crc kubenswrapper[4948]: I0220 08:31:38.055305 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-c08d-account-create-update-54vb7"] Feb 20 08:31:38 crc kubenswrapper[4948]: I0220 08:31:38.067434 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-fs9rr"] Feb 20 08:31:38 crc kubenswrapper[4948]: I0220 08:31:38.084195 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-c08d-account-create-update-54vb7"] Feb 20 08:31:38 crc kubenswrapper[4948]: I0220 08:31:38.167879 4948 generic.go:334] "Generic (PLEG): container finished" podID="060989bf-ef06-4060-bc25-915180f42e5b" containerID="7cfda758c2e1d88101427b25367b79d9870e89cc2808d38e69b1e5dfa5780966" exitCode=0 Feb 20 08:31:38 crc kubenswrapper[4948]: I0220 08:31:38.167927 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdphs" event={"ID":"060989bf-ef06-4060-bc25-915180f42e5b","Type":"ContainerDied","Data":"7cfda758c2e1d88101427b25367b79d9870e89cc2808d38e69b1e5dfa5780966"} Feb 20 08:31:39 crc kubenswrapper[4948]: I0220 08:31:39.179496 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdphs" event={"ID":"060989bf-ef06-4060-bc25-915180f42e5b","Type":"ContainerStarted","Data":"37c7e4d25f9127067f12ac87e6ea1ca5dbb8e45ce19e01d5f213b5bb1cd40529"} Feb 20 08:31:39 crc kubenswrapper[4948]: I0220 08:31:39.205194 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bdphs" podStartSLOduration=2.787117993 podStartE2EDuration="5.205172714s" podCreationTimestamp="2026-02-20 08:31:34 +0000 UTC" firstStartedPulling="2026-02-20 08:31:36.152658963 +0000 UTC m=+1545.127153783" lastFinishedPulling="2026-02-20 08:31:38.570713684 +0000 UTC m=+1547.545208504" observedRunningTime="2026-02-20 08:31:39.195878464 +0000 UTC m=+1548.170373284" watchObservedRunningTime="2026-02-20 08:31:39.205172714 +0000 UTC m=+1548.179667544" Feb 20 08:31:39 crc kubenswrapper[4948]: I0220 08:31:39.740178 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="333ba21a-82a9-4f39-9ec5-afb7ae2ccf60" path="/var/lib/kubelet/pods/333ba21a-82a9-4f39-9ec5-afb7ae2ccf60/volumes" Feb 20 08:31:39 crc kubenswrapper[4948]: I0220 08:31:39.741584 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d815209-521c-4c7f-a026-18899832459f" path="/var/lib/kubelet/pods/4d815209-521c-4c7f-a026-18899832459f/volumes" Feb 20 08:31:45 crc kubenswrapper[4948]: I0220 08:31:45.257506 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bdphs" Feb 20 08:31:45 crc kubenswrapper[4948]: I0220 08:31:45.258235 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bdphs" Feb 20 08:31:46 crc kubenswrapper[4948]: I0220 08:31:46.328623 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bdphs" podUID="060989bf-ef06-4060-bc25-915180f42e5b" containerName="registry-server" probeResult="failure" output=< Feb 20 08:31:46 crc kubenswrapper[4948]: timeout: failed to connect service ":50051" within 1s Feb 20 08:31:46 crc kubenswrapper[4948]: > Feb 20 08:31:49 crc kubenswrapper[4948]: I0220 08:31:49.296733 4948 generic.go:334] "Generic (PLEG): container finished" podID="5864a68d-650e-4bcf-b705-619c0f27445b" containerID="bb15461fc1a5607b927c4774ccf8aa52aa2b1f66a0749cf932076cda97280dfb" exitCode=0 Feb 20 08:31:49 crc kubenswrapper[4948]: I0220 08:31:49.296811 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" event={"ID":"5864a68d-650e-4bcf-b705-619c0f27445b","Type":"ContainerDied","Data":"bb15461fc1a5607b927c4774ccf8aa52aa2b1f66a0749cf932076cda97280dfb"} Feb 20 08:31:49 crc kubenswrapper[4948]: I0220 08:31:49.722707 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:31:49 crc kubenswrapper[4948]: E0220 08:31:49.723130 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:31:50 crc kubenswrapper[4948]: I0220 08:31:50.842671 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" Feb 20 08:31:50 crc kubenswrapper[4948]: I0220 08:31:50.920638 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgbgc\" (UniqueName: \"kubernetes.io/projected/5864a68d-650e-4bcf-b705-619c0f27445b-kube-api-access-lgbgc\") pod \"5864a68d-650e-4bcf-b705-619c0f27445b\" (UID: \"5864a68d-650e-4bcf-b705-619c0f27445b\") " Feb 20 08:31:50 crc kubenswrapper[4948]: I0220 08:31:50.921145 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5864a68d-650e-4bcf-b705-619c0f27445b-inventory\") pod \"5864a68d-650e-4bcf-b705-619c0f27445b\" (UID: \"5864a68d-650e-4bcf-b705-619c0f27445b\") " Feb 20 08:31:50 crc kubenswrapper[4948]: I0220 08:31:50.921490 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5864a68d-650e-4bcf-b705-619c0f27445b-ssh-key-openstack-edpm-ipam\") pod \"5864a68d-650e-4bcf-b705-619c0f27445b\" (UID: \"5864a68d-650e-4bcf-b705-619c0f27445b\") " Feb 20 08:31:50 crc kubenswrapper[4948]: I0220 08:31:50.926098 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5864a68d-650e-4bcf-b705-619c0f27445b-kube-api-access-lgbgc" (OuterVolumeSpecName: "kube-api-access-lgbgc") pod "5864a68d-650e-4bcf-b705-619c0f27445b" (UID: "5864a68d-650e-4bcf-b705-619c0f27445b"). InnerVolumeSpecName "kube-api-access-lgbgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:31:50 crc kubenswrapper[4948]: I0220 08:31:50.951064 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5864a68d-650e-4bcf-b705-619c0f27445b-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5864a68d-650e-4bcf-b705-619c0f27445b" (UID: "5864a68d-650e-4bcf-b705-619c0f27445b"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:31:50 crc kubenswrapper[4948]: I0220 08:31:50.952303 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5864a68d-650e-4bcf-b705-619c0f27445b-inventory" (OuterVolumeSpecName: "inventory") pod "5864a68d-650e-4bcf-b705-619c0f27445b" (UID: "5864a68d-650e-4bcf-b705-619c0f27445b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.024410 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgbgc\" (UniqueName: \"kubernetes.io/projected/5864a68d-650e-4bcf-b705-619c0f27445b-kube-api-access-lgbgc\") on node \"crc\" DevicePath \"\"" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.024440 4948 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5864a68d-650e-4bcf-b705-619c0f27445b-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.024450 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5864a68d-650e-4bcf-b705-619c0f27445b-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.318437 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" event={"ID":"5864a68d-650e-4bcf-b705-619c0f27445b","Type":"ContainerDied","Data":"3ff23193cbb1aafb8b5d373010ecf579e6350e550af86a1ff2fe999d6f3221b3"} Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.318782 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ff23193cbb1aafb8b5d373010ecf579e6350e550af86a1ff2fe999d6f3221b3" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.318529 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.432496 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx"] Feb 20 08:31:51 crc kubenswrapper[4948]: E0220 08:31:51.433015 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5864a68d-650e-4bcf-b705-619c0f27445b" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.433038 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="5864a68d-650e-4bcf-b705-619c0f27445b" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.433301 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="5864a68d-650e-4bcf-b705-619c0f27445b" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.434402 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.439547 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.439847 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.440035 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sgx49" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.440155 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx"] Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.440198 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.534929 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bf3501bf-8f1e-4529-91a8-6be83eda4158-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx\" (UID: \"bf3501bf-8f1e-4529-91a8-6be83eda4158\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.535141 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnztp\" (UniqueName: \"kubernetes.io/projected/bf3501bf-8f1e-4529-91a8-6be83eda4158-kube-api-access-qnztp\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx\" (UID: \"bf3501bf-8f1e-4529-91a8-6be83eda4158\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.535214 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf3501bf-8f1e-4529-91a8-6be83eda4158-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx\" (UID: \"bf3501bf-8f1e-4529-91a8-6be83eda4158\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.637891 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnztp\" (UniqueName: \"kubernetes.io/projected/bf3501bf-8f1e-4529-91a8-6be83eda4158-kube-api-access-qnztp\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx\" (UID: \"bf3501bf-8f1e-4529-91a8-6be83eda4158\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.638013 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf3501bf-8f1e-4529-91a8-6be83eda4158-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx\" (UID: \"bf3501bf-8f1e-4529-91a8-6be83eda4158\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.638306 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bf3501bf-8f1e-4529-91a8-6be83eda4158-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx\" (UID: \"bf3501bf-8f1e-4529-91a8-6be83eda4158\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.642410 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf3501bf-8f1e-4529-91a8-6be83eda4158-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx\" (UID: \"bf3501bf-8f1e-4529-91a8-6be83eda4158\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.645026 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bf3501bf-8f1e-4529-91a8-6be83eda4158-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx\" (UID: \"bf3501bf-8f1e-4529-91a8-6be83eda4158\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.656659 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnztp\" (UniqueName: \"kubernetes.io/projected/bf3501bf-8f1e-4529-91a8-6be83eda4158-kube-api-access-qnztp\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx\" (UID: \"bf3501bf-8f1e-4529-91a8-6be83eda4158\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.760594 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sgx49" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.769411 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" Feb 20 08:31:52 crc kubenswrapper[4948]: I0220 08:31:52.354118 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx"] Feb 20 08:31:52 crc kubenswrapper[4948]: I0220 08:31:52.816709 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:31:53 crc kubenswrapper[4948]: I0220 08:31:53.335211 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" event={"ID":"bf3501bf-8f1e-4529-91a8-6be83eda4158","Type":"ContainerStarted","Data":"91d95d8f8184e4a5f35617935d58a284d64d7340d9543fd4f2b969143f1e8c7b"} Feb 20 08:31:53 crc kubenswrapper[4948]: I0220 08:31:53.335306 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" event={"ID":"bf3501bf-8f1e-4529-91a8-6be83eda4158","Type":"ContainerStarted","Data":"64a64af0744255f035f713b739bbbc9dbd942f32b093f55924592b6aeb394758"} Feb 20 08:31:53 crc kubenswrapper[4948]: I0220 08:31:53.350343 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" podStartSLOduration=1.899962839 podStartE2EDuration="2.350326625s" podCreationTimestamp="2026-02-20 08:31:51 +0000 UTC" firstStartedPulling="2026-02-20 08:31:52.360450283 +0000 UTC m=+1561.334945103" lastFinishedPulling="2026-02-20 08:31:52.810814029 +0000 UTC m=+1561.785308889" observedRunningTime="2026-02-20 08:31:53.349018173 +0000 UTC m=+1562.323513003" watchObservedRunningTime="2026-02-20 08:31:53.350326625 +0000 UTC m=+1562.324821445" Feb 20 08:31:55 crc kubenswrapper[4948]: I0220 08:31:55.345838 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bdphs" Feb 20 08:31:55 crc kubenswrapper[4948]: I0220 08:31:55.438814 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bdphs" Feb 20 08:31:55 crc kubenswrapper[4948]: I0220 08:31:55.599752 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bdphs"] Feb 20 08:31:57 crc kubenswrapper[4948]: I0220 08:31:57.383663 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bdphs" podUID="060989bf-ef06-4060-bc25-915180f42e5b" containerName="registry-server" containerID="cri-o://37c7e4d25f9127067f12ac87e6ea1ca5dbb8e45ce19e01d5f213b5bb1cd40529" gracePeriod=2 Feb 20 08:31:57 crc kubenswrapper[4948]: I0220 08:31:57.954585 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bdphs" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.072844 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4l4qt\" (UniqueName: \"kubernetes.io/projected/060989bf-ef06-4060-bc25-915180f42e5b-kube-api-access-4l4qt\") pod \"060989bf-ef06-4060-bc25-915180f42e5b\" (UID: \"060989bf-ef06-4060-bc25-915180f42e5b\") " Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.073099 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/060989bf-ef06-4060-bc25-915180f42e5b-catalog-content\") pod \"060989bf-ef06-4060-bc25-915180f42e5b\" (UID: \"060989bf-ef06-4060-bc25-915180f42e5b\") " Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.073248 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/060989bf-ef06-4060-bc25-915180f42e5b-utilities\") pod \"060989bf-ef06-4060-bc25-915180f42e5b\" (UID: \"060989bf-ef06-4060-bc25-915180f42e5b\") " Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.074683 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/060989bf-ef06-4060-bc25-915180f42e5b-utilities" (OuterVolumeSpecName: "utilities") pod "060989bf-ef06-4060-bc25-915180f42e5b" (UID: "060989bf-ef06-4060-bc25-915180f42e5b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.087958 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/060989bf-ef06-4060-bc25-915180f42e5b-kube-api-access-4l4qt" (OuterVolumeSpecName: "kube-api-access-4l4qt") pod "060989bf-ef06-4060-bc25-915180f42e5b" (UID: "060989bf-ef06-4060-bc25-915180f42e5b"). InnerVolumeSpecName "kube-api-access-4l4qt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.174948 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/060989bf-ef06-4060-bc25-915180f42e5b-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.175237 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4l4qt\" (UniqueName: \"kubernetes.io/projected/060989bf-ef06-4060-bc25-915180f42e5b-kube-api-access-4l4qt\") on node \"crc\" DevicePath \"\"" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.191165 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/060989bf-ef06-4060-bc25-915180f42e5b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "060989bf-ef06-4060-bc25-915180f42e5b" (UID: "060989bf-ef06-4060-bc25-915180f42e5b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.276646 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/060989bf-ef06-4060-bc25-915180f42e5b-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.396524 4948 generic.go:334] "Generic (PLEG): container finished" podID="060989bf-ef06-4060-bc25-915180f42e5b" containerID="37c7e4d25f9127067f12ac87e6ea1ca5dbb8e45ce19e01d5f213b5bb1cd40529" exitCode=0 Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.396650 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bdphs" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.401113 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdphs" event={"ID":"060989bf-ef06-4060-bc25-915180f42e5b","Type":"ContainerDied","Data":"37c7e4d25f9127067f12ac87e6ea1ca5dbb8e45ce19e01d5f213b5bb1cd40529"} Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.401208 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdphs" event={"ID":"060989bf-ef06-4060-bc25-915180f42e5b","Type":"ContainerDied","Data":"53735cceaab8db6cbf509cf3d7a338fd07617ca1b94b6981218603729cf70256"} Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.401242 4948 scope.go:117] "RemoveContainer" containerID="37c7e4d25f9127067f12ac87e6ea1ca5dbb8e45ce19e01d5f213b5bb1cd40529" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.423202 4948 scope.go:117] "RemoveContainer" containerID="7cfda758c2e1d88101427b25367b79d9870e89cc2808d38e69b1e5dfa5780966" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.440024 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bdphs"] Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.449474 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bdphs"] Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.467923 4948 scope.go:117] "RemoveContainer" containerID="cafef04c966377e665f78102a5914bb6441b7f707970965b08dc70cfb2cf82a1" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.539144 4948 scope.go:117] "RemoveContainer" containerID="37c7e4d25f9127067f12ac87e6ea1ca5dbb8e45ce19e01d5f213b5bb1cd40529" Feb 20 08:31:58 crc kubenswrapper[4948]: E0220 08:31:58.540360 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37c7e4d25f9127067f12ac87e6ea1ca5dbb8e45ce19e01d5f213b5bb1cd40529\": container with ID starting with 37c7e4d25f9127067f12ac87e6ea1ca5dbb8e45ce19e01d5f213b5bb1cd40529 not found: ID does not exist" containerID="37c7e4d25f9127067f12ac87e6ea1ca5dbb8e45ce19e01d5f213b5bb1cd40529" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.540390 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37c7e4d25f9127067f12ac87e6ea1ca5dbb8e45ce19e01d5f213b5bb1cd40529"} err="failed to get container status \"37c7e4d25f9127067f12ac87e6ea1ca5dbb8e45ce19e01d5f213b5bb1cd40529\": rpc error: code = NotFound desc = could not find container \"37c7e4d25f9127067f12ac87e6ea1ca5dbb8e45ce19e01d5f213b5bb1cd40529\": container with ID starting with 37c7e4d25f9127067f12ac87e6ea1ca5dbb8e45ce19e01d5f213b5bb1cd40529 not found: ID does not exist" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.540410 4948 scope.go:117] "RemoveContainer" containerID="7cfda758c2e1d88101427b25367b79d9870e89cc2808d38e69b1e5dfa5780966" Feb 20 08:31:58 crc kubenswrapper[4948]: E0220 08:31:58.540695 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7cfda758c2e1d88101427b25367b79d9870e89cc2808d38e69b1e5dfa5780966\": container with ID starting with 7cfda758c2e1d88101427b25367b79d9870e89cc2808d38e69b1e5dfa5780966 not found: ID does not exist" containerID="7cfda758c2e1d88101427b25367b79d9870e89cc2808d38e69b1e5dfa5780966" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.540715 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cfda758c2e1d88101427b25367b79d9870e89cc2808d38e69b1e5dfa5780966"} err="failed to get container status \"7cfda758c2e1d88101427b25367b79d9870e89cc2808d38e69b1e5dfa5780966\": rpc error: code = NotFound desc = could not find container \"7cfda758c2e1d88101427b25367b79d9870e89cc2808d38e69b1e5dfa5780966\": container with ID starting with 7cfda758c2e1d88101427b25367b79d9870e89cc2808d38e69b1e5dfa5780966 not found: ID does not exist" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.540727 4948 scope.go:117] "RemoveContainer" containerID="cafef04c966377e665f78102a5914bb6441b7f707970965b08dc70cfb2cf82a1" Feb 20 08:31:58 crc kubenswrapper[4948]: E0220 08:31:58.540990 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cafef04c966377e665f78102a5914bb6441b7f707970965b08dc70cfb2cf82a1\": container with ID starting with cafef04c966377e665f78102a5914bb6441b7f707970965b08dc70cfb2cf82a1 not found: ID does not exist" containerID="cafef04c966377e665f78102a5914bb6441b7f707970965b08dc70cfb2cf82a1" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.541010 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cafef04c966377e665f78102a5914bb6441b7f707970965b08dc70cfb2cf82a1"} err="failed to get container status \"cafef04c966377e665f78102a5914bb6441b7f707970965b08dc70cfb2cf82a1\": rpc error: code = NotFound desc = could not find container \"cafef04c966377e665f78102a5914bb6441b7f707970965b08dc70cfb2cf82a1\": container with ID starting with cafef04c966377e665f78102a5914bb6441b7f707970965b08dc70cfb2cf82a1 not found: ID does not exist" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.887000 4948 scope.go:117] "RemoveContainer" containerID="f4eb6ffc3a340bf3fb020244f7c41526c57d80da4f478c0885a3916041b2da30" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.915904 4948 scope.go:117] "RemoveContainer" containerID="67a7241e5369f13541e0b84997b90ee0fd96a87f376c9c419e0e637fa48068da" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.972338 4948 scope.go:117] "RemoveContainer" containerID="279cb01f22f256d8185916f133561b9dbb69d1fb0ef3e531faed5ee14470b9d2" Feb 20 08:31:59 crc kubenswrapper[4948]: I0220 08:31:59.034929 4948 scope.go:117] "RemoveContainer" containerID="043fb8ebd535abc3ac64988e423e5b76f657b01785d57b024ae113afd44b4bf7" Feb 20 08:31:59 crc kubenswrapper[4948]: I0220 08:31:59.083922 4948 scope.go:117] "RemoveContainer" containerID="eb75184d7c11178b70689bbef778a20eca59576d9cf6e2826963a0598aa904a4" Feb 20 08:31:59 crc kubenswrapper[4948]: I0220 08:31:59.116349 4948 scope.go:117] "RemoveContainer" containerID="8e0c0f098f3973e04052c6dc5c4da1ff436370579d9b1ea6857d9e0c6bed683e" Feb 20 08:31:59 crc kubenswrapper[4948]: I0220 08:31:59.748548 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="060989bf-ef06-4060-bc25-915180f42e5b" path="/var/lib/kubelet/pods/060989bf-ef06-4060-bc25-915180f42e5b/volumes" Feb 20 08:32:00 crc kubenswrapper[4948]: I0220 08:32:00.044180 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-a86b-account-create-update-mspp7"] Feb 20 08:32:00 crc kubenswrapper[4948]: I0220 08:32:00.064553 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-97j89"] Feb 20 08:32:00 crc kubenswrapper[4948]: I0220 08:32:00.075931 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-4djkr"] Feb 20 08:32:00 crc kubenswrapper[4948]: I0220 08:32:00.086152 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-a86b-account-create-update-mspp7"] Feb 20 08:32:00 crc kubenswrapper[4948]: I0220 08:32:00.096064 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-97j89"] Feb 20 08:32:00 crc kubenswrapper[4948]: I0220 08:32:00.102904 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-4djkr"] Feb 20 08:32:01 crc kubenswrapper[4948]: I0220 08:32:01.047766 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-e6d0-account-create-update-cdxd5"] Feb 20 08:32:01 crc kubenswrapper[4948]: I0220 08:32:01.070049 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-w4bzp"] Feb 20 08:32:01 crc kubenswrapper[4948]: I0220 08:32:01.089314 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-w4bzp"] Feb 20 08:32:01 crc kubenswrapper[4948]: I0220 08:32:01.099364 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-e6d0-account-create-update-cdxd5"] Feb 20 08:32:01 crc kubenswrapper[4948]: I0220 08:32:01.107785 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-0b2f-account-create-update-2f74q"] Feb 20 08:32:01 crc kubenswrapper[4948]: I0220 08:32:01.115484 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-zvzvk"] Feb 20 08:32:01 crc kubenswrapper[4948]: I0220 08:32:01.123401 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-zvzvk"] Feb 20 08:32:01 crc kubenswrapper[4948]: I0220 08:32:01.130817 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-0b2f-account-create-update-2f74q"] Feb 20 08:32:01 crc kubenswrapper[4948]: I0220 08:32:01.743731 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a739e4d-8960-46f3-8c5d-2373b59165dd" path="/var/lib/kubelet/pods/0a739e4d-8960-46f3-8c5d-2373b59165dd/volumes" Feb 20 08:32:01 crc kubenswrapper[4948]: I0220 08:32:01.746533 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10eedd01-3f52-4d44-981d-27bc16af186b" path="/var/lib/kubelet/pods/10eedd01-3f52-4d44-981d-27bc16af186b/volumes" Feb 20 08:32:01 crc kubenswrapper[4948]: I0220 08:32:01.748126 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12f1f665-7a04-4794-a727-cc5c060a1933" path="/var/lib/kubelet/pods/12f1f665-7a04-4794-a727-cc5c060a1933/volumes" Feb 20 08:32:01 crc kubenswrapper[4948]: I0220 08:32:01.749568 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ced4ac5-e057-476a-908a-f573a8cf997b" path="/var/lib/kubelet/pods/4ced4ac5-e057-476a-908a-f573a8cf997b/volumes" Feb 20 08:32:01 crc kubenswrapper[4948]: I0220 08:32:01.752240 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="501e504c-7910-47cb-a181-812a0d72f2db" path="/var/lib/kubelet/pods/501e504c-7910-47cb-a181-812a0d72f2db/volumes" Feb 20 08:32:01 crc kubenswrapper[4948]: I0220 08:32:01.754700 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6d068f3-210b-41e8-b90a-5406759c0606" path="/var/lib/kubelet/pods/b6d068f3-210b-41e8-b90a-5406759c0606/volumes" Feb 20 08:32:01 crc kubenswrapper[4948]: I0220 08:32:01.756880 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3f86085-113f-408e-b43c-a499e2fe8b90" path="/var/lib/kubelet/pods/d3f86085-113f-408e-b43c-a499e2fe8b90/volumes" Feb 20 08:32:04 crc kubenswrapper[4948]: I0220 08:32:04.724532 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:32:04 crc kubenswrapper[4948]: E0220 08:32:04.725418 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:32:12 crc kubenswrapper[4948]: I0220 08:32:12.063107 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-pkgzx"] Feb 20 08:32:12 crc kubenswrapper[4948]: I0220 08:32:12.077308 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-pkgzx"] Feb 20 08:32:13 crc kubenswrapper[4948]: I0220 08:32:13.036154 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-gz5j7"] Feb 20 08:32:13 crc kubenswrapper[4948]: I0220 08:32:13.047571 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-gz5j7"] Feb 20 08:32:13 crc kubenswrapper[4948]: I0220 08:32:13.756751 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65735728-e734-40de-a185-9caa1db9a47d" path="/var/lib/kubelet/pods/65735728-e734-40de-a185-9caa1db9a47d/volumes" Feb 20 08:32:13 crc kubenswrapper[4948]: I0220 08:32:13.759151 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9370de8-8b72-4819-af79-4a9ee1fb6777" path="/var/lib/kubelet/pods/b9370de8-8b72-4819-af79-4a9ee1fb6777/volumes" Feb 20 08:32:17 crc kubenswrapper[4948]: I0220 08:32:17.723505 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:32:17 crc kubenswrapper[4948]: E0220 08:32:17.724689 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:32:30 crc kubenswrapper[4948]: I0220 08:32:30.723247 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:32:30 crc kubenswrapper[4948]: E0220 08:32:30.724285 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:32:41 crc kubenswrapper[4948]: I0220 08:32:41.060853 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-g5zxh"] Feb 20 08:32:41 crc kubenswrapper[4948]: I0220 08:32:41.073044 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-g5zxh"] Feb 20 08:32:41 crc kubenswrapper[4948]: I0220 08:32:41.745863 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb362d0e-68b4-4d48-885c-a6894ac71e6e" path="/var/lib/kubelet/pods/bb362d0e-68b4-4d48-885c-a6894ac71e6e/volumes" Feb 20 08:32:43 crc kubenswrapper[4948]: I0220 08:32:43.723862 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:32:43 crc kubenswrapper[4948]: E0220 08:32:43.724478 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:32:43 crc kubenswrapper[4948]: I0220 08:32:43.959617 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wfjj9"] Feb 20 08:32:43 crc kubenswrapper[4948]: E0220 08:32:43.960297 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="060989bf-ef06-4060-bc25-915180f42e5b" containerName="registry-server" Feb 20 08:32:43 crc kubenswrapper[4948]: I0220 08:32:43.960329 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="060989bf-ef06-4060-bc25-915180f42e5b" containerName="registry-server" Feb 20 08:32:43 crc kubenswrapper[4948]: E0220 08:32:43.960404 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="060989bf-ef06-4060-bc25-915180f42e5b" containerName="extract-utilities" Feb 20 08:32:43 crc kubenswrapper[4948]: I0220 08:32:43.960419 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="060989bf-ef06-4060-bc25-915180f42e5b" containerName="extract-utilities" Feb 20 08:32:43 crc kubenswrapper[4948]: E0220 08:32:43.960451 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="060989bf-ef06-4060-bc25-915180f42e5b" containerName="extract-content" Feb 20 08:32:43 crc kubenswrapper[4948]: I0220 08:32:43.960464 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="060989bf-ef06-4060-bc25-915180f42e5b" containerName="extract-content" Feb 20 08:32:43 crc kubenswrapper[4948]: I0220 08:32:43.960812 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="060989bf-ef06-4060-bc25-915180f42e5b" containerName="registry-server" Feb 20 08:32:43 crc kubenswrapper[4948]: I0220 08:32:43.963553 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wfjj9" Feb 20 08:32:43 crc kubenswrapper[4948]: I0220 08:32:43.976177 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wfjj9"] Feb 20 08:32:44 crc kubenswrapper[4948]: I0220 08:32:44.038045 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ng6rg\" (UniqueName: \"kubernetes.io/projected/127d52c1-3c58-424b-aa24-d71a40715228-kube-api-access-ng6rg\") pod \"certified-operators-wfjj9\" (UID: \"127d52c1-3c58-424b-aa24-d71a40715228\") " pod="openshift-marketplace/certified-operators-wfjj9" Feb 20 08:32:44 crc kubenswrapper[4948]: I0220 08:32:44.038143 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/127d52c1-3c58-424b-aa24-d71a40715228-catalog-content\") pod \"certified-operators-wfjj9\" (UID: \"127d52c1-3c58-424b-aa24-d71a40715228\") " pod="openshift-marketplace/certified-operators-wfjj9" Feb 20 08:32:44 crc kubenswrapper[4948]: I0220 08:32:44.038185 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/127d52c1-3c58-424b-aa24-d71a40715228-utilities\") pod \"certified-operators-wfjj9\" (UID: \"127d52c1-3c58-424b-aa24-d71a40715228\") " pod="openshift-marketplace/certified-operators-wfjj9" Feb 20 08:32:44 crc kubenswrapper[4948]: I0220 08:32:44.140508 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ng6rg\" (UniqueName: \"kubernetes.io/projected/127d52c1-3c58-424b-aa24-d71a40715228-kube-api-access-ng6rg\") pod \"certified-operators-wfjj9\" (UID: \"127d52c1-3c58-424b-aa24-d71a40715228\") " pod="openshift-marketplace/certified-operators-wfjj9" Feb 20 08:32:44 crc kubenswrapper[4948]: I0220 08:32:44.141110 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/127d52c1-3c58-424b-aa24-d71a40715228-catalog-content\") pod \"certified-operators-wfjj9\" (UID: \"127d52c1-3c58-424b-aa24-d71a40715228\") " pod="openshift-marketplace/certified-operators-wfjj9" Feb 20 08:32:44 crc kubenswrapper[4948]: I0220 08:32:44.141647 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/127d52c1-3c58-424b-aa24-d71a40715228-catalog-content\") pod \"certified-operators-wfjj9\" (UID: \"127d52c1-3c58-424b-aa24-d71a40715228\") " pod="openshift-marketplace/certified-operators-wfjj9" Feb 20 08:32:44 crc kubenswrapper[4948]: I0220 08:32:44.141808 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/127d52c1-3c58-424b-aa24-d71a40715228-utilities\") pod \"certified-operators-wfjj9\" (UID: \"127d52c1-3c58-424b-aa24-d71a40715228\") " pod="openshift-marketplace/certified-operators-wfjj9" Feb 20 08:32:44 crc kubenswrapper[4948]: I0220 08:32:44.142180 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/127d52c1-3c58-424b-aa24-d71a40715228-utilities\") pod \"certified-operators-wfjj9\" (UID: \"127d52c1-3c58-424b-aa24-d71a40715228\") " pod="openshift-marketplace/certified-operators-wfjj9" Feb 20 08:32:44 crc kubenswrapper[4948]: I0220 08:32:44.162953 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ng6rg\" (UniqueName: \"kubernetes.io/projected/127d52c1-3c58-424b-aa24-d71a40715228-kube-api-access-ng6rg\") pod \"certified-operators-wfjj9\" (UID: \"127d52c1-3c58-424b-aa24-d71a40715228\") " pod="openshift-marketplace/certified-operators-wfjj9" Feb 20 08:32:44 crc kubenswrapper[4948]: I0220 08:32:44.298192 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wfjj9" Feb 20 08:32:44 crc kubenswrapper[4948]: I0220 08:32:44.792859 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wfjj9"] Feb 20 08:32:45 crc kubenswrapper[4948]: I0220 08:32:45.009183 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfjj9" event={"ID":"127d52c1-3c58-424b-aa24-d71a40715228","Type":"ContainerStarted","Data":"105e4d157fc4d190d06fdac23834273f0c6149fd0638550ef4ca602d493f78ed"} Feb 20 08:32:45 crc kubenswrapper[4948]: I0220 08:32:45.009501 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfjj9" event={"ID":"127d52c1-3c58-424b-aa24-d71a40715228","Type":"ContainerStarted","Data":"162bd0680e1ceb7df91470f556615acedd676fe4c2560f55f65606b3bd82a0ad"} Feb 20 08:32:46 crc kubenswrapper[4948]: I0220 08:32:46.025432 4948 generic.go:334] "Generic (PLEG): container finished" podID="127d52c1-3c58-424b-aa24-d71a40715228" containerID="105e4d157fc4d190d06fdac23834273f0c6149fd0638550ef4ca602d493f78ed" exitCode=0 Feb 20 08:32:46 crc kubenswrapper[4948]: I0220 08:32:46.025506 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfjj9" event={"ID":"127d52c1-3c58-424b-aa24-d71a40715228","Type":"ContainerDied","Data":"105e4d157fc4d190d06fdac23834273f0c6149fd0638550ef4ca602d493f78ed"} Feb 20 08:32:46 crc kubenswrapper[4948]: I0220 08:32:46.029149 4948 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 08:32:47 crc kubenswrapper[4948]: I0220 08:32:47.030189 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-6mz7x"] Feb 20 08:32:47 crc kubenswrapper[4948]: I0220 08:32:47.038077 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfjj9" event={"ID":"127d52c1-3c58-424b-aa24-d71a40715228","Type":"ContainerStarted","Data":"88f25f8239c60da55713318b991058f1de64af08883114ae420d1004b33150e5"} Feb 20 08:32:47 crc kubenswrapper[4948]: I0220 08:32:47.039242 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-6mz7x"] Feb 20 08:32:47 crc kubenswrapper[4948]: I0220 08:32:47.747463 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1227b752-8a3c-4c01-91bf-0662c1b1a231" path="/var/lib/kubelet/pods/1227b752-8a3c-4c01-91bf-0662c1b1a231/volumes" Feb 20 08:32:48 crc kubenswrapper[4948]: I0220 08:32:48.053294 4948 generic.go:334] "Generic (PLEG): container finished" podID="127d52c1-3c58-424b-aa24-d71a40715228" containerID="88f25f8239c60da55713318b991058f1de64af08883114ae420d1004b33150e5" exitCode=0 Feb 20 08:32:48 crc kubenswrapper[4948]: I0220 08:32:48.053358 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfjj9" event={"ID":"127d52c1-3c58-424b-aa24-d71a40715228","Type":"ContainerDied","Data":"88f25f8239c60da55713318b991058f1de64af08883114ae420d1004b33150e5"} Feb 20 08:32:49 crc kubenswrapper[4948]: I0220 08:32:49.066244 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfjj9" event={"ID":"127d52c1-3c58-424b-aa24-d71a40715228","Type":"ContainerStarted","Data":"058114f0c0b1e6905b0749e5f435db738c2b0ddf4ed180217b4ddd8152ed9e31"} Feb 20 08:32:49 crc kubenswrapper[4948]: I0220 08:32:49.106419 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wfjj9" podStartSLOduration=3.695804652 podStartE2EDuration="6.106395068s" podCreationTimestamp="2026-02-20 08:32:43 +0000 UTC" firstStartedPulling="2026-02-20 08:32:46.028561749 +0000 UTC m=+1615.003056599" lastFinishedPulling="2026-02-20 08:32:48.439152195 +0000 UTC m=+1617.413647015" observedRunningTime="2026-02-20 08:32:49.099582089 +0000 UTC m=+1618.074076909" watchObservedRunningTime="2026-02-20 08:32:49.106395068 +0000 UTC m=+1618.080889908" Feb 20 08:32:50 crc kubenswrapper[4948]: I0220 08:32:50.061325 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-v29fx"] Feb 20 08:32:50 crc kubenswrapper[4948]: I0220 08:32:50.077529 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-v29fx"] Feb 20 08:32:51 crc kubenswrapper[4948]: I0220 08:32:51.738726 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d84a71b6-7503-4150-a5c4-4579b08a669a" path="/var/lib/kubelet/pods/d84a71b6-7503-4150-a5c4-4579b08a669a/volumes" Feb 20 08:32:54 crc kubenswrapper[4948]: I0220 08:32:54.300461 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wfjj9" Feb 20 08:32:54 crc kubenswrapper[4948]: I0220 08:32:54.300545 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wfjj9" Feb 20 08:32:54 crc kubenswrapper[4948]: I0220 08:32:54.392448 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wfjj9" Feb 20 08:32:55 crc kubenswrapper[4948]: I0220 08:32:55.228922 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wfjj9" Feb 20 08:32:55 crc kubenswrapper[4948]: I0220 08:32:55.309610 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wfjj9"] Feb 20 08:32:57 crc kubenswrapper[4948]: I0220 08:32:57.166778 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wfjj9" podUID="127d52c1-3c58-424b-aa24-d71a40715228" containerName="registry-server" containerID="cri-o://058114f0c0b1e6905b0749e5f435db738c2b0ddf4ed180217b4ddd8152ed9e31" gracePeriod=2 Feb 20 08:32:57 crc kubenswrapper[4948]: I0220 08:32:57.723142 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:32:57 crc kubenswrapper[4948]: E0220 08:32:57.723749 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:32:57 crc kubenswrapper[4948]: I0220 08:32:57.735128 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wfjj9" Feb 20 08:32:57 crc kubenswrapper[4948]: I0220 08:32:57.854092 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ng6rg\" (UniqueName: \"kubernetes.io/projected/127d52c1-3c58-424b-aa24-d71a40715228-kube-api-access-ng6rg\") pod \"127d52c1-3c58-424b-aa24-d71a40715228\" (UID: \"127d52c1-3c58-424b-aa24-d71a40715228\") " Feb 20 08:32:57 crc kubenswrapper[4948]: I0220 08:32:57.854494 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/127d52c1-3c58-424b-aa24-d71a40715228-catalog-content\") pod \"127d52c1-3c58-424b-aa24-d71a40715228\" (UID: \"127d52c1-3c58-424b-aa24-d71a40715228\") " Feb 20 08:32:57 crc kubenswrapper[4948]: I0220 08:32:57.854599 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/127d52c1-3c58-424b-aa24-d71a40715228-utilities\") pod \"127d52c1-3c58-424b-aa24-d71a40715228\" (UID: \"127d52c1-3c58-424b-aa24-d71a40715228\") " Feb 20 08:32:57 crc kubenswrapper[4948]: I0220 08:32:57.857961 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/127d52c1-3c58-424b-aa24-d71a40715228-utilities" (OuterVolumeSpecName: "utilities") pod "127d52c1-3c58-424b-aa24-d71a40715228" (UID: "127d52c1-3c58-424b-aa24-d71a40715228"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:32:57 crc kubenswrapper[4948]: I0220 08:32:57.864053 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/127d52c1-3c58-424b-aa24-d71a40715228-kube-api-access-ng6rg" (OuterVolumeSpecName: "kube-api-access-ng6rg") pod "127d52c1-3c58-424b-aa24-d71a40715228" (UID: "127d52c1-3c58-424b-aa24-d71a40715228"). InnerVolumeSpecName "kube-api-access-ng6rg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:32:57 crc kubenswrapper[4948]: I0220 08:32:57.960689 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/127d52c1-3c58-424b-aa24-d71a40715228-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:32:57 crc kubenswrapper[4948]: I0220 08:32:57.960749 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ng6rg\" (UniqueName: \"kubernetes.io/projected/127d52c1-3c58-424b-aa24-d71a40715228-kube-api-access-ng6rg\") on node \"crc\" DevicePath \"\"" Feb 20 08:32:58 crc kubenswrapper[4948]: I0220 08:32:58.010920 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/127d52c1-3c58-424b-aa24-d71a40715228-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "127d52c1-3c58-424b-aa24-d71a40715228" (UID: "127d52c1-3c58-424b-aa24-d71a40715228"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:32:58 crc kubenswrapper[4948]: I0220 08:32:58.062304 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/127d52c1-3c58-424b-aa24-d71a40715228-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:32:58 crc kubenswrapper[4948]: I0220 08:32:58.179830 4948 generic.go:334] "Generic (PLEG): container finished" podID="127d52c1-3c58-424b-aa24-d71a40715228" containerID="058114f0c0b1e6905b0749e5f435db738c2b0ddf4ed180217b4ddd8152ed9e31" exitCode=0 Feb 20 08:32:58 crc kubenswrapper[4948]: I0220 08:32:58.179881 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfjj9" event={"ID":"127d52c1-3c58-424b-aa24-d71a40715228","Type":"ContainerDied","Data":"058114f0c0b1e6905b0749e5f435db738c2b0ddf4ed180217b4ddd8152ed9e31"} Feb 20 08:32:58 crc kubenswrapper[4948]: I0220 08:32:58.179912 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfjj9" event={"ID":"127d52c1-3c58-424b-aa24-d71a40715228","Type":"ContainerDied","Data":"162bd0680e1ceb7df91470f556615acedd676fe4c2560f55f65606b3bd82a0ad"} Feb 20 08:32:58 crc kubenswrapper[4948]: I0220 08:32:58.179934 4948 scope.go:117] "RemoveContainer" containerID="058114f0c0b1e6905b0749e5f435db738c2b0ddf4ed180217b4ddd8152ed9e31" Feb 20 08:32:58 crc kubenswrapper[4948]: I0220 08:32:58.180144 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wfjj9" Feb 20 08:32:58 crc kubenswrapper[4948]: I0220 08:32:58.228109 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wfjj9"] Feb 20 08:32:58 crc kubenswrapper[4948]: I0220 08:32:58.236537 4948 scope.go:117] "RemoveContainer" containerID="88f25f8239c60da55713318b991058f1de64af08883114ae420d1004b33150e5" Feb 20 08:32:58 crc kubenswrapper[4948]: I0220 08:32:58.242662 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wfjj9"] Feb 20 08:32:58 crc kubenswrapper[4948]: I0220 08:32:58.274530 4948 scope.go:117] "RemoveContainer" containerID="105e4d157fc4d190d06fdac23834273f0c6149fd0638550ef4ca602d493f78ed" Feb 20 08:32:58 crc kubenswrapper[4948]: I0220 08:32:58.342643 4948 scope.go:117] "RemoveContainer" containerID="058114f0c0b1e6905b0749e5f435db738c2b0ddf4ed180217b4ddd8152ed9e31" Feb 20 08:32:58 crc kubenswrapper[4948]: E0220 08:32:58.343234 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"058114f0c0b1e6905b0749e5f435db738c2b0ddf4ed180217b4ddd8152ed9e31\": container with ID starting with 058114f0c0b1e6905b0749e5f435db738c2b0ddf4ed180217b4ddd8152ed9e31 not found: ID does not exist" containerID="058114f0c0b1e6905b0749e5f435db738c2b0ddf4ed180217b4ddd8152ed9e31" Feb 20 08:32:58 crc kubenswrapper[4948]: I0220 08:32:58.343276 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"058114f0c0b1e6905b0749e5f435db738c2b0ddf4ed180217b4ddd8152ed9e31"} err="failed to get container status \"058114f0c0b1e6905b0749e5f435db738c2b0ddf4ed180217b4ddd8152ed9e31\": rpc error: code = NotFound desc = could not find container \"058114f0c0b1e6905b0749e5f435db738c2b0ddf4ed180217b4ddd8152ed9e31\": container with ID starting with 058114f0c0b1e6905b0749e5f435db738c2b0ddf4ed180217b4ddd8152ed9e31 not found: ID does not exist" Feb 20 08:32:58 crc kubenswrapper[4948]: I0220 08:32:58.343302 4948 scope.go:117] "RemoveContainer" containerID="88f25f8239c60da55713318b991058f1de64af08883114ae420d1004b33150e5" Feb 20 08:32:58 crc kubenswrapper[4948]: E0220 08:32:58.343720 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88f25f8239c60da55713318b991058f1de64af08883114ae420d1004b33150e5\": container with ID starting with 88f25f8239c60da55713318b991058f1de64af08883114ae420d1004b33150e5 not found: ID does not exist" containerID="88f25f8239c60da55713318b991058f1de64af08883114ae420d1004b33150e5" Feb 20 08:32:58 crc kubenswrapper[4948]: I0220 08:32:58.343887 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88f25f8239c60da55713318b991058f1de64af08883114ae420d1004b33150e5"} err="failed to get container status \"88f25f8239c60da55713318b991058f1de64af08883114ae420d1004b33150e5\": rpc error: code = NotFound desc = could not find container \"88f25f8239c60da55713318b991058f1de64af08883114ae420d1004b33150e5\": container with ID starting with 88f25f8239c60da55713318b991058f1de64af08883114ae420d1004b33150e5 not found: ID does not exist" Feb 20 08:32:58 crc kubenswrapper[4948]: I0220 08:32:58.344023 4948 scope.go:117] "RemoveContainer" containerID="105e4d157fc4d190d06fdac23834273f0c6149fd0638550ef4ca602d493f78ed" Feb 20 08:32:58 crc kubenswrapper[4948]: E0220 08:32:58.344458 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"105e4d157fc4d190d06fdac23834273f0c6149fd0638550ef4ca602d493f78ed\": container with ID starting with 105e4d157fc4d190d06fdac23834273f0c6149fd0638550ef4ca602d493f78ed not found: ID does not exist" containerID="105e4d157fc4d190d06fdac23834273f0c6149fd0638550ef4ca602d493f78ed" Feb 20 08:32:58 crc kubenswrapper[4948]: I0220 08:32:58.344576 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"105e4d157fc4d190d06fdac23834273f0c6149fd0638550ef4ca602d493f78ed"} err="failed to get container status \"105e4d157fc4d190d06fdac23834273f0c6149fd0638550ef4ca602d493f78ed\": rpc error: code = NotFound desc = could not find container \"105e4d157fc4d190d06fdac23834273f0c6149fd0638550ef4ca602d493f78ed\": container with ID starting with 105e4d157fc4d190d06fdac23834273f0c6149fd0638550ef4ca602d493f78ed not found: ID does not exist" Feb 20 08:32:59 crc kubenswrapper[4948]: I0220 08:32:59.197647 4948 generic.go:334] "Generic (PLEG): container finished" podID="bf3501bf-8f1e-4529-91a8-6be83eda4158" containerID="91d95d8f8184e4a5f35617935d58a284d64d7340d9543fd4f2b969143f1e8c7b" exitCode=0 Feb 20 08:32:59 crc kubenswrapper[4948]: I0220 08:32:59.197733 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" event={"ID":"bf3501bf-8f1e-4529-91a8-6be83eda4158","Type":"ContainerDied","Data":"91d95d8f8184e4a5f35617935d58a284d64d7340d9543fd4f2b969143f1e8c7b"} Feb 20 08:32:59 crc kubenswrapper[4948]: I0220 08:32:59.326685 4948 scope.go:117] "RemoveContainer" containerID="6457842f65c42b8313a38ca922ee61a97a4823ed4c56040981289a5c5705cd46" Feb 20 08:32:59 crc kubenswrapper[4948]: I0220 08:32:59.365367 4948 scope.go:117] "RemoveContainer" containerID="f8ab1b0def6ea5b1022acdb9c87659d2c1c20a66a736d1e630b89d097c36d6ef" Feb 20 08:32:59 crc kubenswrapper[4948]: I0220 08:32:59.438707 4948 scope.go:117] "RemoveContainer" containerID="87c2cdb1a2b7f5c38a01dd7f884ce5938563a33612ebaf51f20ceac5532f97f2" Feb 20 08:32:59 crc kubenswrapper[4948]: I0220 08:32:59.519780 4948 scope.go:117] "RemoveContainer" containerID="fb257235a0241c571d44f0c21ce5fac86cd6393e8a1250cb5e78d4e1b599fb51" Feb 20 08:32:59 crc kubenswrapper[4948]: I0220 08:32:59.553943 4948 scope.go:117] "RemoveContainer" containerID="1fb571973e057c8519a5db05840734117d94ed09a1e521d57cf4a92fa88cbcfb" Feb 20 08:32:59 crc kubenswrapper[4948]: I0220 08:32:59.614812 4948 scope.go:117] "RemoveContainer" containerID="7e3cc540d00613674fe548789ab2ccc71a06c5714f420cedea97409a8254b14f" Feb 20 08:32:59 crc kubenswrapper[4948]: I0220 08:32:59.640400 4948 scope.go:117] "RemoveContainer" containerID="c565785e94684cef9e257c737e709e95f61b9c29c0e4f1d8e37358207acddb4c" Feb 20 08:32:59 crc kubenswrapper[4948]: I0220 08:32:59.672123 4948 scope.go:117] "RemoveContainer" containerID="b8d5cd74c5cbd29dae34c7b1cedbde475f451d2de997a59b511db6cc61ba0a29" Feb 20 08:32:59 crc kubenswrapper[4948]: I0220 08:32:59.697613 4948 scope.go:117] "RemoveContainer" containerID="cddaaa0c08e72fa702328c57550853b538c4537573467ac0ebd0761e1fcbc30c" Feb 20 08:32:59 crc kubenswrapper[4948]: I0220 08:32:59.720067 4948 scope.go:117] "RemoveContainer" containerID="cb7241154041475439806eb631212532adb6e6775adf70e62bdf5cf4b6d19276" Feb 20 08:32:59 crc kubenswrapper[4948]: I0220 08:32:59.741746 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="127d52c1-3c58-424b-aa24-d71a40715228" path="/var/lib/kubelet/pods/127d52c1-3c58-424b-aa24-d71a40715228/volumes" Feb 20 08:32:59 crc kubenswrapper[4948]: I0220 08:32:59.757837 4948 scope.go:117] "RemoveContainer" containerID="72b50822f4c48e3166b0066730c258d864d6094e98d9d9617ca0abdc1f36ae38" Feb 20 08:32:59 crc kubenswrapper[4948]: I0220 08:32:59.797425 4948 scope.go:117] "RemoveContainer" containerID="a11f29a4f832ccc8dd04f52fcb05e1f418934455b1143a9d46f2d5ad1bce9425" Feb 20 08:32:59 crc kubenswrapper[4948]: I0220 08:32:59.826276 4948 scope.go:117] "RemoveContainer" containerID="ede14cf7eee35beb81bccaf96c5abbd270f76ba98b65d67579bbb482b130f45e" Feb 20 08:32:59 crc kubenswrapper[4948]: I0220 08:32:59.847415 4948 scope.go:117] "RemoveContainer" containerID="ddcd74111702fbb6556b1046c04ec320ea3aecaa381a53da7a6aa1bc0c3d6fd7" Feb 20 08:33:00 crc kubenswrapper[4948]: I0220 08:33:00.860103 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" Feb 20 08:33:00 crc kubenswrapper[4948]: I0220 08:33:00.924819 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf3501bf-8f1e-4529-91a8-6be83eda4158-inventory\") pod \"bf3501bf-8f1e-4529-91a8-6be83eda4158\" (UID: \"bf3501bf-8f1e-4529-91a8-6be83eda4158\") " Feb 20 08:33:00 crc kubenswrapper[4948]: I0220 08:33:00.925046 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnztp\" (UniqueName: \"kubernetes.io/projected/bf3501bf-8f1e-4529-91a8-6be83eda4158-kube-api-access-qnztp\") pod \"bf3501bf-8f1e-4529-91a8-6be83eda4158\" (UID: \"bf3501bf-8f1e-4529-91a8-6be83eda4158\") " Feb 20 08:33:00 crc kubenswrapper[4948]: I0220 08:33:00.925344 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bf3501bf-8f1e-4529-91a8-6be83eda4158-ssh-key-openstack-edpm-ipam\") pod \"bf3501bf-8f1e-4529-91a8-6be83eda4158\" (UID: \"bf3501bf-8f1e-4529-91a8-6be83eda4158\") " Feb 20 08:33:00 crc kubenswrapper[4948]: I0220 08:33:00.930925 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf3501bf-8f1e-4529-91a8-6be83eda4158-kube-api-access-qnztp" (OuterVolumeSpecName: "kube-api-access-qnztp") pod "bf3501bf-8f1e-4529-91a8-6be83eda4158" (UID: "bf3501bf-8f1e-4529-91a8-6be83eda4158"). InnerVolumeSpecName "kube-api-access-qnztp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:33:00 crc kubenswrapper[4948]: I0220 08:33:00.955695 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf3501bf-8f1e-4529-91a8-6be83eda4158-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "bf3501bf-8f1e-4529-91a8-6be83eda4158" (UID: "bf3501bf-8f1e-4529-91a8-6be83eda4158"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:33:00 crc kubenswrapper[4948]: I0220 08:33:00.971417 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf3501bf-8f1e-4529-91a8-6be83eda4158-inventory" (OuterVolumeSpecName: "inventory") pod "bf3501bf-8f1e-4529-91a8-6be83eda4158" (UID: "bf3501bf-8f1e-4529-91a8-6be83eda4158"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.027820 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bf3501bf-8f1e-4529-91a8-6be83eda4158-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.027875 4948 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf3501bf-8f1e-4529-91a8-6be83eda4158-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.027895 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnztp\" (UniqueName: \"kubernetes.io/projected/bf3501bf-8f1e-4529-91a8-6be83eda4158-kube-api-access-qnztp\") on node \"crc\" DevicePath \"\"" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.238906 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" event={"ID":"bf3501bf-8f1e-4529-91a8-6be83eda4158","Type":"ContainerDied","Data":"64a64af0744255f035f713b739bbbc9dbd942f32b093f55924592b6aeb394758"} Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.240180 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64a64af0744255f035f713b739bbbc9dbd942f32b093f55924592b6aeb394758" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.239002 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.542434 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d"] Feb 20 08:33:01 crc kubenswrapper[4948]: E0220 08:33:01.542799 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="127d52c1-3c58-424b-aa24-d71a40715228" containerName="extract-utilities" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.542816 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="127d52c1-3c58-424b-aa24-d71a40715228" containerName="extract-utilities" Feb 20 08:33:01 crc kubenswrapper[4948]: E0220 08:33:01.542833 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf3501bf-8f1e-4529-91a8-6be83eda4158" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.542841 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf3501bf-8f1e-4529-91a8-6be83eda4158" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 20 08:33:01 crc kubenswrapper[4948]: E0220 08:33:01.542866 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="127d52c1-3c58-424b-aa24-d71a40715228" containerName="extract-content" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.542872 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="127d52c1-3c58-424b-aa24-d71a40715228" containerName="extract-content" Feb 20 08:33:01 crc kubenswrapper[4948]: E0220 08:33:01.542884 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="127d52c1-3c58-424b-aa24-d71a40715228" containerName="registry-server" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.542890 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="127d52c1-3c58-424b-aa24-d71a40715228" containerName="registry-server" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.543079 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="127d52c1-3c58-424b-aa24-d71a40715228" containerName="registry-server" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.543094 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf3501bf-8f1e-4529-91a8-6be83eda4158" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.543705 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.546852 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.546864 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.547532 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sgx49" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.551795 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.574374 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d"] Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.639823 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lf98d\" (UID: \"2e05be38-f4ab-415f-a71d-9e5233cf0ea7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.639902 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lf98d\" (UID: \"2e05be38-f4ab-415f-a71d-9e5233cf0ea7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.640297 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfwmf\" (UniqueName: \"kubernetes.io/projected/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-kube-api-access-vfwmf\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lf98d\" (UID: \"2e05be38-f4ab-415f-a71d-9e5233cf0ea7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.742176 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lf98d\" (UID: \"2e05be38-f4ab-415f-a71d-9e5233cf0ea7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.742264 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lf98d\" (UID: \"2e05be38-f4ab-415f-a71d-9e5233cf0ea7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.742462 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfwmf\" (UniqueName: \"kubernetes.io/projected/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-kube-api-access-vfwmf\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lf98d\" (UID: \"2e05be38-f4ab-415f-a71d-9e5233cf0ea7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.749431 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lf98d\" (UID: \"2e05be38-f4ab-415f-a71d-9e5233cf0ea7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.750035 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lf98d\" (UID: \"2e05be38-f4ab-415f-a71d-9e5233cf0ea7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.772884 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfwmf\" (UniqueName: \"kubernetes.io/projected/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-kube-api-access-vfwmf\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lf98d\" (UID: \"2e05be38-f4ab-415f-a71d-9e5233cf0ea7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.863514 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" Feb 20 08:33:02 crc kubenswrapper[4948]: I0220 08:33:02.424340 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d"] Feb 20 08:33:03 crc kubenswrapper[4948]: I0220 08:33:03.264466 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" event={"ID":"2e05be38-f4ab-415f-a71d-9e5233cf0ea7","Type":"ContainerStarted","Data":"345677fd825b78a8c9a15733dc5504746aec837d4a4da61dd8b843aa56c4878f"} Feb 20 08:33:03 crc kubenswrapper[4948]: I0220 08:33:03.265091 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" event={"ID":"2e05be38-f4ab-415f-a71d-9e5233cf0ea7","Type":"ContainerStarted","Data":"3eb7fcb6fef9877434c717228e6f27c38c4af180ee08b3677022845e3cf3f71d"} Feb 20 08:33:03 crc kubenswrapper[4948]: I0220 08:33:03.291785 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" podStartSLOduration=1.8341930990000002 podStartE2EDuration="2.291766439s" podCreationTimestamp="2026-02-20 08:33:01 +0000 UTC" firstStartedPulling="2026-02-20 08:33:02.439579229 +0000 UTC m=+1631.414074049" lastFinishedPulling="2026-02-20 08:33:02.897152529 +0000 UTC m=+1631.871647389" observedRunningTime="2026-02-20 08:33:03.281892685 +0000 UTC m=+1632.256387525" watchObservedRunningTime="2026-02-20 08:33:03.291766439 +0000 UTC m=+1632.266261259" Feb 20 08:33:04 crc kubenswrapper[4948]: I0220 08:33:04.050397 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-xt2zq"] Feb 20 08:33:04 crc kubenswrapper[4948]: I0220 08:33:04.061342 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-xt2zq"] Feb 20 08:33:05 crc kubenswrapper[4948]: I0220 08:33:05.738442 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f706b741-86f9-49a3-95d5-85ee62eb3668" path="/var/lib/kubelet/pods/f706b741-86f9-49a3-95d5-85ee62eb3668/volumes" Feb 20 08:33:06 crc kubenswrapper[4948]: I0220 08:33:06.028657 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-c7zzv"] Feb 20 08:33:06 crc kubenswrapper[4948]: I0220 08:33:06.042626 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-c7zzv"] Feb 20 08:33:07 crc kubenswrapper[4948]: I0220 08:33:07.746089 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1f108e1-27bb-433e-b7e1-f9cc7f778182" path="/var/lib/kubelet/pods/f1f108e1-27bb-433e-b7e1-f9cc7f778182/volumes" Feb 20 08:33:08 crc kubenswrapper[4948]: I0220 08:33:08.337501 4948 generic.go:334] "Generic (PLEG): container finished" podID="2e05be38-f4ab-415f-a71d-9e5233cf0ea7" containerID="345677fd825b78a8c9a15733dc5504746aec837d4a4da61dd8b843aa56c4878f" exitCode=0 Feb 20 08:33:08 crc kubenswrapper[4948]: I0220 08:33:08.337625 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" event={"ID":"2e05be38-f4ab-415f-a71d-9e5233cf0ea7","Type":"ContainerDied","Data":"345677fd825b78a8c9a15733dc5504746aec837d4a4da61dd8b843aa56c4878f"} Feb 20 08:33:09 crc kubenswrapper[4948]: I0220 08:33:09.882849 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.065247 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-inventory\") pod \"2e05be38-f4ab-415f-a71d-9e5233cf0ea7\" (UID: \"2e05be38-f4ab-415f-a71d-9e5233cf0ea7\") " Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.065509 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-ssh-key-openstack-edpm-ipam\") pod \"2e05be38-f4ab-415f-a71d-9e5233cf0ea7\" (UID: \"2e05be38-f4ab-415f-a71d-9e5233cf0ea7\") " Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.065576 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfwmf\" (UniqueName: \"kubernetes.io/projected/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-kube-api-access-vfwmf\") pod \"2e05be38-f4ab-415f-a71d-9e5233cf0ea7\" (UID: \"2e05be38-f4ab-415f-a71d-9e5233cf0ea7\") " Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.073768 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-kube-api-access-vfwmf" (OuterVolumeSpecName: "kube-api-access-vfwmf") pod "2e05be38-f4ab-415f-a71d-9e5233cf0ea7" (UID: "2e05be38-f4ab-415f-a71d-9e5233cf0ea7"). InnerVolumeSpecName "kube-api-access-vfwmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.094400 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-inventory" (OuterVolumeSpecName: "inventory") pod "2e05be38-f4ab-415f-a71d-9e5233cf0ea7" (UID: "2e05be38-f4ab-415f-a71d-9e5233cf0ea7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.106385 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "2e05be38-f4ab-415f-a71d-9e5233cf0ea7" (UID: "2e05be38-f4ab-415f-a71d-9e5233cf0ea7"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.168203 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.168242 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfwmf\" (UniqueName: \"kubernetes.io/projected/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-kube-api-access-vfwmf\") on node \"crc\" DevicePath \"\"" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.168256 4948 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.367167 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" event={"ID":"2e05be38-f4ab-415f-a71d-9e5233cf0ea7","Type":"ContainerDied","Data":"3eb7fcb6fef9877434c717228e6f27c38c4af180ee08b3677022845e3cf3f71d"} Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.367226 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3eb7fcb6fef9877434c717228e6f27c38c4af180ee08b3677022845e3cf3f71d" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.367351 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.586490 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v"] Feb 20 08:33:10 crc kubenswrapper[4948]: E0220 08:33:10.588413 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e05be38-f4ab-415f-a71d-9e5233cf0ea7" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.588465 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e05be38-f4ab-415f-a71d-9e5233cf0ea7" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.589558 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e05be38-f4ab-415f-a71d-9e5233cf0ea7" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.593538 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.600283 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sgx49" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.600367 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.600397 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.601167 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.611849 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v"] Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.677414 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f42ddd0c-a2ec-414a-812d-f87213a3226d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4l55v\" (UID: \"f42ddd0c-a2ec-414a-812d-f87213a3226d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.677753 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxqrc\" (UniqueName: \"kubernetes.io/projected/f42ddd0c-a2ec-414a-812d-f87213a3226d-kube-api-access-dxqrc\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4l55v\" (UID: \"f42ddd0c-a2ec-414a-812d-f87213a3226d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.677850 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f42ddd0c-a2ec-414a-812d-f87213a3226d-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4l55v\" (UID: \"f42ddd0c-a2ec-414a-812d-f87213a3226d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.779586 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f42ddd0c-a2ec-414a-812d-f87213a3226d-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4l55v\" (UID: \"f42ddd0c-a2ec-414a-812d-f87213a3226d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.779714 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f42ddd0c-a2ec-414a-812d-f87213a3226d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4l55v\" (UID: \"f42ddd0c-a2ec-414a-812d-f87213a3226d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.779843 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxqrc\" (UniqueName: \"kubernetes.io/projected/f42ddd0c-a2ec-414a-812d-f87213a3226d-kube-api-access-dxqrc\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4l55v\" (UID: \"f42ddd0c-a2ec-414a-812d-f87213a3226d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.785589 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f42ddd0c-a2ec-414a-812d-f87213a3226d-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4l55v\" (UID: \"f42ddd0c-a2ec-414a-812d-f87213a3226d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.786901 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f42ddd0c-a2ec-414a-812d-f87213a3226d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4l55v\" (UID: \"f42ddd0c-a2ec-414a-812d-f87213a3226d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.805213 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxqrc\" (UniqueName: \"kubernetes.io/projected/f42ddd0c-a2ec-414a-812d-f87213a3226d-kube-api-access-dxqrc\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4l55v\" (UID: \"f42ddd0c-a2ec-414a-812d-f87213a3226d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.931970 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" Feb 20 08:33:11 crc kubenswrapper[4948]: I0220 08:33:11.501608 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v"] Feb 20 08:33:11 crc kubenswrapper[4948]: I0220 08:33:11.729047 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:33:11 crc kubenswrapper[4948]: E0220 08:33:11.729345 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:33:12 crc kubenswrapper[4948]: I0220 08:33:12.393255 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" event={"ID":"f42ddd0c-a2ec-414a-812d-f87213a3226d","Type":"ContainerStarted","Data":"ddcb74ec6513e9050b354f5584993e70b4784d82329c695b8cbcea156ddada02"} Feb 20 08:33:12 crc kubenswrapper[4948]: I0220 08:33:12.394009 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" event={"ID":"f42ddd0c-a2ec-414a-812d-f87213a3226d","Type":"ContainerStarted","Data":"26e225a9996f58185c76eebc01d4b0d42e12e26c4c12d0fd5527a74efcfa0e7f"} Feb 20 08:33:12 crc kubenswrapper[4948]: I0220 08:33:12.420865 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" podStartSLOduration=2.015293287 podStartE2EDuration="2.420844738s" podCreationTimestamp="2026-02-20 08:33:10 +0000 UTC" firstStartedPulling="2026-02-20 08:33:11.50705241 +0000 UTC m=+1640.481547230" lastFinishedPulling="2026-02-20 08:33:11.912603851 +0000 UTC m=+1640.887098681" observedRunningTime="2026-02-20 08:33:12.419385541 +0000 UTC m=+1641.393880361" watchObservedRunningTime="2026-02-20 08:33:12.420844738 +0000 UTC m=+1641.395339558" Feb 20 08:33:22 crc kubenswrapper[4948]: I0220 08:33:22.722757 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:33:22 crc kubenswrapper[4948]: E0220 08:33:22.725554 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:33:23 crc kubenswrapper[4948]: I0220 08:33:23.564806 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lvjzf"] Feb 20 08:33:23 crc kubenswrapper[4948]: I0220 08:33:23.567279 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lvjzf" Feb 20 08:33:23 crc kubenswrapper[4948]: I0220 08:33:23.600761 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lvjzf"] Feb 20 08:33:23 crc kubenswrapper[4948]: I0220 08:33:23.667489 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jfnl\" (UniqueName: \"kubernetes.io/projected/26bee9f7-36e7-4452-be46-a7af14fe0d82-kube-api-access-2jfnl\") pod \"community-operators-lvjzf\" (UID: \"26bee9f7-36e7-4452-be46-a7af14fe0d82\") " pod="openshift-marketplace/community-operators-lvjzf" Feb 20 08:33:23 crc kubenswrapper[4948]: I0220 08:33:23.667765 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26bee9f7-36e7-4452-be46-a7af14fe0d82-utilities\") pod \"community-operators-lvjzf\" (UID: \"26bee9f7-36e7-4452-be46-a7af14fe0d82\") " pod="openshift-marketplace/community-operators-lvjzf" Feb 20 08:33:23 crc kubenswrapper[4948]: I0220 08:33:23.667937 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26bee9f7-36e7-4452-be46-a7af14fe0d82-catalog-content\") pod \"community-operators-lvjzf\" (UID: \"26bee9f7-36e7-4452-be46-a7af14fe0d82\") " pod="openshift-marketplace/community-operators-lvjzf" Feb 20 08:33:23 crc kubenswrapper[4948]: I0220 08:33:23.769407 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jfnl\" (UniqueName: \"kubernetes.io/projected/26bee9f7-36e7-4452-be46-a7af14fe0d82-kube-api-access-2jfnl\") pod \"community-operators-lvjzf\" (UID: \"26bee9f7-36e7-4452-be46-a7af14fe0d82\") " pod="openshift-marketplace/community-operators-lvjzf" Feb 20 08:33:23 crc kubenswrapper[4948]: I0220 08:33:23.769540 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26bee9f7-36e7-4452-be46-a7af14fe0d82-utilities\") pod \"community-operators-lvjzf\" (UID: \"26bee9f7-36e7-4452-be46-a7af14fe0d82\") " pod="openshift-marketplace/community-operators-lvjzf" Feb 20 08:33:23 crc kubenswrapper[4948]: I0220 08:33:23.769619 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26bee9f7-36e7-4452-be46-a7af14fe0d82-catalog-content\") pod \"community-operators-lvjzf\" (UID: \"26bee9f7-36e7-4452-be46-a7af14fe0d82\") " pod="openshift-marketplace/community-operators-lvjzf" Feb 20 08:33:23 crc kubenswrapper[4948]: I0220 08:33:23.770092 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26bee9f7-36e7-4452-be46-a7af14fe0d82-catalog-content\") pod \"community-operators-lvjzf\" (UID: \"26bee9f7-36e7-4452-be46-a7af14fe0d82\") " pod="openshift-marketplace/community-operators-lvjzf" Feb 20 08:33:23 crc kubenswrapper[4948]: I0220 08:33:23.770090 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26bee9f7-36e7-4452-be46-a7af14fe0d82-utilities\") pod \"community-operators-lvjzf\" (UID: \"26bee9f7-36e7-4452-be46-a7af14fe0d82\") " pod="openshift-marketplace/community-operators-lvjzf" Feb 20 08:33:23 crc kubenswrapper[4948]: I0220 08:33:23.798173 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jfnl\" (UniqueName: \"kubernetes.io/projected/26bee9f7-36e7-4452-be46-a7af14fe0d82-kube-api-access-2jfnl\") pod \"community-operators-lvjzf\" (UID: \"26bee9f7-36e7-4452-be46-a7af14fe0d82\") " pod="openshift-marketplace/community-operators-lvjzf" Feb 20 08:33:23 crc kubenswrapper[4948]: I0220 08:33:23.894405 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lvjzf" Feb 20 08:33:24 crc kubenswrapper[4948]: I0220 08:33:24.417039 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lvjzf"] Feb 20 08:33:24 crc kubenswrapper[4948]: I0220 08:33:24.544625 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lvjzf" event={"ID":"26bee9f7-36e7-4452-be46-a7af14fe0d82","Type":"ContainerStarted","Data":"d87e55975b4c132e628eaad19e343e2dc1a36778a825ecd32677e5386de184a8"} Feb 20 08:33:24 crc kubenswrapper[4948]: E0220 08:33:24.823961 4948 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod26bee9f7_36e7_4452_be46_a7af14fe0d82.slice/crio-conmon-d1d54224a90227b551f9ae7bab7701dda7f4f8e05b155c101b1a90b05f80f77a.scope\": RecentStats: unable to find data in memory cache]" Feb 20 08:33:25 crc kubenswrapper[4948]: I0220 08:33:25.559497 4948 generic.go:334] "Generic (PLEG): container finished" podID="26bee9f7-36e7-4452-be46-a7af14fe0d82" containerID="d1d54224a90227b551f9ae7bab7701dda7f4f8e05b155c101b1a90b05f80f77a" exitCode=0 Feb 20 08:33:25 crc kubenswrapper[4948]: I0220 08:33:25.559621 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lvjzf" event={"ID":"26bee9f7-36e7-4452-be46-a7af14fe0d82","Type":"ContainerDied","Data":"d1d54224a90227b551f9ae7bab7701dda7f4f8e05b155c101b1a90b05f80f77a"} Feb 20 08:33:26 crc kubenswrapper[4948]: I0220 08:33:26.572659 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lvjzf" event={"ID":"26bee9f7-36e7-4452-be46-a7af14fe0d82","Type":"ContainerStarted","Data":"3bd508455cdc61cc2aed3667a9786384d7d8905d6c881e37112bbb71d52d5a9c"} Feb 20 08:33:27 crc kubenswrapper[4948]: I0220 08:33:27.582706 4948 generic.go:334] "Generic (PLEG): container finished" podID="26bee9f7-36e7-4452-be46-a7af14fe0d82" containerID="3bd508455cdc61cc2aed3667a9786384d7d8905d6c881e37112bbb71d52d5a9c" exitCode=0 Feb 20 08:33:27 crc kubenswrapper[4948]: I0220 08:33:27.582745 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lvjzf" event={"ID":"26bee9f7-36e7-4452-be46-a7af14fe0d82","Type":"ContainerDied","Data":"3bd508455cdc61cc2aed3667a9786384d7d8905d6c881e37112bbb71d52d5a9c"} Feb 20 08:33:28 crc kubenswrapper[4948]: I0220 08:33:28.593448 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lvjzf" event={"ID":"26bee9f7-36e7-4452-be46-a7af14fe0d82","Type":"ContainerStarted","Data":"2ed5eb1d692e2406cfd129f46d82525b4996228bb8988814eef3b67af71f3d1d"} Feb 20 08:33:28 crc kubenswrapper[4948]: I0220 08:33:28.612331 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lvjzf" podStartSLOduration=3.184383818 podStartE2EDuration="5.612310293s" podCreationTimestamp="2026-02-20 08:33:23 +0000 UTC" firstStartedPulling="2026-02-20 08:33:25.562612168 +0000 UTC m=+1654.537106988" lastFinishedPulling="2026-02-20 08:33:27.990538633 +0000 UTC m=+1656.965033463" observedRunningTime="2026-02-20 08:33:28.611033452 +0000 UTC m=+1657.585528272" watchObservedRunningTime="2026-02-20 08:33:28.612310293 +0000 UTC m=+1657.586805123" Feb 20 08:33:33 crc kubenswrapper[4948]: I0220 08:33:33.895660 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lvjzf" Feb 20 08:33:33 crc kubenswrapper[4948]: I0220 08:33:33.896338 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lvjzf" Feb 20 08:33:33 crc kubenswrapper[4948]: I0220 08:33:33.980385 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lvjzf" Feb 20 08:33:34 crc kubenswrapper[4948]: I0220 08:33:34.744601 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lvjzf" Feb 20 08:33:34 crc kubenswrapper[4948]: I0220 08:33:34.820609 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lvjzf"] Feb 20 08:33:36 crc kubenswrapper[4948]: I0220 08:33:36.674646 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lvjzf" podUID="26bee9f7-36e7-4452-be46-a7af14fe0d82" containerName="registry-server" containerID="cri-o://2ed5eb1d692e2406cfd129f46d82525b4996228bb8988814eef3b67af71f3d1d" gracePeriod=2 Feb 20 08:33:36 crc kubenswrapper[4948]: I0220 08:33:36.722776 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:33:36 crc kubenswrapper[4948]: E0220 08:33:36.723047 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.191211 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lvjzf" Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.285607 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26bee9f7-36e7-4452-be46-a7af14fe0d82-utilities\") pod \"26bee9f7-36e7-4452-be46-a7af14fe0d82\" (UID: \"26bee9f7-36e7-4452-be46-a7af14fe0d82\") " Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.285659 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jfnl\" (UniqueName: \"kubernetes.io/projected/26bee9f7-36e7-4452-be46-a7af14fe0d82-kube-api-access-2jfnl\") pod \"26bee9f7-36e7-4452-be46-a7af14fe0d82\" (UID: \"26bee9f7-36e7-4452-be46-a7af14fe0d82\") " Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.285787 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26bee9f7-36e7-4452-be46-a7af14fe0d82-catalog-content\") pod \"26bee9f7-36e7-4452-be46-a7af14fe0d82\" (UID: \"26bee9f7-36e7-4452-be46-a7af14fe0d82\") " Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.286748 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26bee9f7-36e7-4452-be46-a7af14fe0d82-utilities" (OuterVolumeSpecName: "utilities") pod "26bee9f7-36e7-4452-be46-a7af14fe0d82" (UID: "26bee9f7-36e7-4452-be46-a7af14fe0d82"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.292173 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26bee9f7-36e7-4452-be46-a7af14fe0d82-kube-api-access-2jfnl" (OuterVolumeSpecName: "kube-api-access-2jfnl") pod "26bee9f7-36e7-4452-be46-a7af14fe0d82" (UID: "26bee9f7-36e7-4452-be46-a7af14fe0d82"). InnerVolumeSpecName "kube-api-access-2jfnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.335618 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26bee9f7-36e7-4452-be46-a7af14fe0d82-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "26bee9f7-36e7-4452-be46-a7af14fe0d82" (UID: "26bee9f7-36e7-4452-be46-a7af14fe0d82"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.388236 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26bee9f7-36e7-4452-be46-a7af14fe0d82-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.388293 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jfnl\" (UniqueName: \"kubernetes.io/projected/26bee9f7-36e7-4452-be46-a7af14fe0d82-kube-api-access-2jfnl\") on node \"crc\" DevicePath \"\"" Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.388316 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26bee9f7-36e7-4452-be46-a7af14fe0d82-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.684757 4948 generic.go:334] "Generic (PLEG): container finished" podID="26bee9f7-36e7-4452-be46-a7af14fe0d82" containerID="2ed5eb1d692e2406cfd129f46d82525b4996228bb8988814eef3b67af71f3d1d" exitCode=0 Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.684843 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lvjzf" Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.684866 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lvjzf" event={"ID":"26bee9f7-36e7-4452-be46-a7af14fe0d82","Type":"ContainerDied","Data":"2ed5eb1d692e2406cfd129f46d82525b4996228bb8988814eef3b67af71f3d1d"} Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.685214 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lvjzf" event={"ID":"26bee9f7-36e7-4452-be46-a7af14fe0d82","Type":"ContainerDied","Data":"d87e55975b4c132e628eaad19e343e2dc1a36778a825ecd32677e5386de184a8"} Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.685282 4948 scope.go:117] "RemoveContainer" containerID="2ed5eb1d692e2406cfd129f46d82525b4996228bb8988814eef3b67af71f3d1d" Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.721260 4948 scope.go:117] "RemoveContainer" containerID="3bd508455cdc61cc2aed3667a9786384d7d8905d6c881e37112bbb71d52d5a9c" Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.747899 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lvjzf"] Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.756417 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lvjzf"] Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.758509 4948 scope.go:117] "RemoveContainer" containerID="d1d54224a90227b551f9ae7bab7701dda7f4f8e05b155c101b1a90b05f80f77a" Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.818181 4948 scope.go:117] "RemoveContainer" containerID="2ed5eb1d692e2406cfd129f46d82525b4996228bb8988814eef3b67af71f3d1d" Feb 20 08:33:37 crc kubenswrapper[4948]: E0220 08:33:37.818738 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ed5eb1d692e2406cfd129f46d82525b4996228bb8988814eef3b67af71f3d1d\": container with ID starting with 2ed5eb1d692e2406cfd129f46d82525b4996228bb8988814eef3b67af71f3d1d not found: ID does not exist" containerID="2ed5eb1d692e2406cfd129f46d82525b4996228bb8988814eef3b67af71f3d1d" Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.818776 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ed5eb1d692e2406cfd129f46d82525b4996228bb8988814eef3b67af71f3d1d"} err="failed to get container status \"2ed5eb1d692e2406cfd129f46d82525b4996228bb8988814eef3b67af71f3d1d\": rpc error: code = NotFound desc = could not find container \"2ed5eb1d692e2406cfd129f46d82525b4996228bb8988814eef3b67af71f3d1d\": container with ID starting with 2ed5eb1d692e2406cfd129f46d82525b4996228bb8988814eef3b67af71f3d1d not found: ID does not exist" Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.818799 4948 scope.go:117] "RemoveContainer" containerID="3bd508455cdc61cc2aed3667a9786384d7d8905d6c881e37112bbb71d52d5a9c" Feb 20 08:33:37 crc kubenswrapper[4948]: E0220 08:33:37.819071 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bd508455cdc61cc2aed3667a9786384d7d8905d6c881e37112bbb71d52d5a9c\": container with ID starting with 3bd508455cdc61cc2aed3667a9786384d7d8905d6c881e37112bbb71d52d5a9c not found: ID does not exist" containerID="3bd508455cdc61cc2aed3667a9786384d7d8905d6c881e37112bbb71d52d5a9c" Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.819099 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bd508455cdc61cc2aed3667a9786384d7d8905d6c881e37112bbb71d52d5a9c"} err="failed to get container status \"3bd508455cdc61cc2aed3667a9786384d7d8905d6c881e37112bbb71d52d5a9c\": rpc error: code = NotFound desc = could not find container \"3bd508455cdc61cc2aed3667a9786384d7d8905d6c881e37112bbb71d52d5a9c\": container with ID starting with 3bd508455cdc61cc2aed3667a9786384d7d8905d6c881e37112bbb71d52d5a9c not found: ID does not exist" Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.819118 4948 scope.go:117] "RemoveContainer" containerID="d1d54224a90227b551f9ae7bab7701dda7f4f8e05b155c101b1a90b05f80f77a" Feb 20 08:33:37 crc kubenswrapper[4948]: E0220 08:33:37.819368 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1d54224a90227b551f9ae7bab7701dda7f4f8e05b155c101b1a90b05f80f77a\": container with ID starting with d1d54224a90227b551f9ae7bab7701dda7f4f8e05b155c101b1a90b05f80f77a not found: ID does not exist" containerID="d1d54224a90227b551f9ae7bab7701dda7f4f8e05b155c101b1a90b05f80f77a" Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.819390 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1d54224a90227b551f9ae7bab7701dda7f4f8e05b155c101b1a90b05f80f77a"} err="failed to get container status \"d1d54224a90227b551f9ae7bab7701dda7f4f8e05b155c101b1a90b05f80f77a\": rpc error: code = NotFound desc = could not find container \"d1d54224a90227b551f9ae7bab7701dda7f4f8e05b155c101b1a90b05f80f77a\": container with ID starting with d1d54224a90227b551f9ae7bab7701dda7f4f8e05b155c101b1a90b05f80f77a not found: ID does not exist" Feb 20 08:33:38 crc kubenswrapper[4948]: I0220 08:33:38.085540 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-aecf-account-create-update-gnnvg"] Feb 20 08:33:38 crc kubenswrapper[4948]: I0220 08:33:38.102603 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-mgnlb"] Feb 20 08:33:38 crc kubenswrapper[4948]: I0220 08:33:38.111189 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-2rgc6"] Feb 20 08:33:38 crc kubenswrapper[4948]: I0220 08:33:38.120715 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-ca33-account-create-update-dwqpd"] Feb 20 08:33:38 crc kubenswrapper[4948]: I0220 08:33:38.130573 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-aecf-account-create-update-gnnvg"] Feb 20 08:33:38 crc kubenswrapper[4948]: I0220 08:33:38.137149 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-2rgc6"] Feb 20 08:33:38 crc kubenswrapper[4948]: I0220 08:33:38.143398 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-mgnlb"] Feb 20 08:33:38 crc kubenswrapper[4948]: I0220 08:33:38.149361 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-ca33-account-create-update-dwqpd"] Feb 20 08:33:39 crc kubenswrapper[4948]: I0220 08:33:39.045603 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-x2pm7"] Feb 20 08:33:39 crc kubenswrapper[4948]: I0220 08:33:39.059173 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-78de-account-create-update-fwjdc"] Feb 20 08:33:39 crc kubenswrapper[4948]: I0220 08:33:39.072535 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-x2pm7"] Feb 20 08:33:39 crc kubenswrapper[4948]: I0220 08:33:39.082966 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-78de-account-create-update-fwjdc"] Feb 20 08:33:39 crc kubenswrapper[4948]: I0220 08:33:39.737454 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b61242-b8c1-4e1b-b319-3eafb79ef048" path="/var/lib/kubelet/pods/20b61242-b8c1-4e1b-b319-3eafb79ef048/volumes" Feb 20 08:33:39 crc kubenswrapper[4948]: I0220 08:33:39.738671 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26bee9f7-36e7-4452-be46-a7af14fe0d82" path="/var/lib/kubelet/pods/26bee9f7-36e7-4452-be46-a7af14fe0d82/volumes" Feb 20 08:33:39 crc kubenswrapper[4948]: I0220 08:33:39.740372 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a311ded-ab56-4f07-8b7f-9dcc4d70f647" path="/var/lib/kubelet/pods/8a311ded-ab56-4f07-8b7f-9dcc4d70f647/volumes" Feb 20 08:33:39 crc kubenswrapper[4948]: I0220 08:33:39.742429 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b167b56-655a-4220-9823-8cc606f5f034" path="/var/lib/kubelet/pods/8b167b56-655a-4220-9823-8cc606f5f034/volumes" Feb 20 08:33:39 crc kubenswrapper[4948]: I0220 08:33:39.743555 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5cb8c15-68fc-47db-9f68-1a8401f99f88" path="/var/lib/kubelet/pods/a5cb8c15-68fc-47db-9f68-1a8401f99f88/volumes" Feb 20 08:33:39 crc kubenswrapper[4948]: I0220 08:33:39.744677 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f" path="/var/lib/kubelet/pods/b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f/volumes" Feb 20 08:33:39 crc kubenswrapper[4948]: I0220 08:33:39.745776 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa398740-9a54-472f-9736-0959dd54b657" path="/var/lib/kubelet/pods/fa398740-9a54-472f-9736-0959dd54b657/volumes" Feb 20 08:33:48 crc kubenswrapper[4948]: I0220 08:33:48.723047 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:33:48 crc kubenswrapper[4948]: E0220 08:33:48.724192 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:33:48 crc kubenswrapper[4948]: I0220 08:33:48.812086 4948 generic.go:334] "Generic (PLEG): container finished" podID="f42ddd0c-a2ec-414a-812d-f87213a3226d" containerID="ddcb74ec6513e9050b354f5584993e70b4784d82329c695b8cbcea156ddada02" exitCode=0 Feb 20 08:33:48 crc kubenswrapper[4948]: I0220 08:33:48.812149 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" event={"ID":"f42ddd0c-a2ec-414a-812d-f87213a3226d","Type":"ContainerDied","Data":"ddcb74ec6513e9050b354f5584993e70b4784d82329c695b8cbcea156ddada02"} Feb 20 08:33:50 crc kubenswrapper[4948]: I0220 08:33:50.354065 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" Feb 20 08:33:50 crc kubenswrapper[4948]: I0220 08:33:50.472490 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f42ddd0c-a2ec-414a-812d-f87213a3226d-inventory\") pod \"f42ddd0c-a2ec-414a-812d-f87213a3226d\" (UID: \"f42ddd0c-a2ec-414a-812d-f87213a3226d\") " Feb 20 08:33:50 crc kubenswrapper[4948]: I0220 08:33:50.472594 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f42ddd0c-a2ec-414a-812d-f87213a3226d-ssh-key-openstack-edpm-ipam\") pod \"f42ddd0c-a2ec-414a-812d-f87213a3226d\" (UID: \"f42ddd0c-a2ec-414a-812d-f87213a3226d\") " Feb 20 08:33:50 crc kubenswrapper[4948]: I0220 08:33:50.472662 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxqrc\" (UniqueName: \"kubernetes.io/projected/f42ddd0c-a2ec-414a-812d-f87213a3226d-kube-api-access-dxqrc\") pod \"f42ddd0c-a2ec-414a-812d-f87213a3226d\" (UID: \"f42ddd0c-a2ec-414a-812d-f87213a3226d\") " Feb 20 08:33:50 crc kubenswrapper[4948]: I0220 08:33:50.478725 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f42ddd0c-a2ec-414a-812d-f87213a3226d-kube-api-access-dxqrc" (OuterVolumeSpecName: "kube-api-access-dxqrc") pod "f42ddd0c-a2ec-414a-812d-f87213a3226d" (UID: "f42ddd0c-a2ec-414a-812d-f87213a3226d"). InnerVolumeSpecName "kube-api-access-dxqrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:33:50 crc kubenswrapper[4948]: I0220 08:33:50.520184 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f42ddd0c-a2ec-414a-812d-f87213a3226d-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "f42ddd0c-a2ec-414a-812d-f87213a3226d" (UID: "f42ddd0c-a2ec-414a-812d-f87213a3226d"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:33:50 crc kubenswrapper[4948]: I0220 08:33:50.531499 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f42ddd0c-a2ec-414a-812d-f87213a3226d-inventory" (OuterVolumeSpecName: "inventory") pod "f42ddd0c-a2ec-414a-812d-f87213a3226d" (UID: "f42ddd0c-a2ec-414a-812d-f87213a3226d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:33:50 crc kubenswrapper[4948]: I0220 08:33:50.575113 4948 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f42ddd0c-a2ec-414a-812d-f87213a3226d-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 08:33:50 crc kubenswrapper[4948]: I0220 08:33:50.575159 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f42ddd0c-a2ec-414a-812d-f87213a3226d-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:33:50 crc kubenswrapper[4948]: I0220 08:33:50.575178 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxqrc\" (UniqueName: \"kubernetes.io/projected/f42ddd0c-a2ec-414a-812d-f87213a3226d-kube-api-access-dxqrc\") on node \"crc\" DevicePath \"\"" Feb 20 08:33:50 crc kubenswrapper[4948]: I0220 08:33:50.835800 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" event={"ID":"f42ddd0c-a2ec-414a-812d-f87213a3226d","Type":"ContainerDied","Data":"26e225a9996f58185c76eebc01d4b0d42e12e26c4c12d0fd5527a74efcfa0e7f"} Feb 20 08:33:50 crc kubenswrapper[4948]: I0220 08:33:50.835861 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="26e225a9996f58185c76eebc01d4b0d42e12e26c4c12d0fd5527a74efcfa0e7f" Feb 20 08:33:50 crc kubenswrapper[4948]: I0220 08:33:50.835937 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" Feb 20 08:33:50 crc kubenswrapper[4948]: I0220 08:33:50.998821 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6"] Feb 20 08:33:50 crc kubenswrapper[4948]: E0220 08:33:50.999806 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26bee9f7-36e7-4452-be46-a7af14fe0d82" containerName="extract-utilities" Feb 20 08:33:50 crc kubenswrapper[4948]: I0220 08:33:50.999836 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="26bee9f7-36e7-4452-be46-a7af14fe0d82" containerName="extract-utilities" Feb 20 08:33:50 crc kubenswrapper[4948]: E0220 08:33:50.999883 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26bee9f7-36e7-4452-be46-a7af14fe0d82" containerName="registry-server" Feb 20 08:33:50 crc kubenswrapper[4948]: I0220 08:33:50.999896 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="26bee9f7-36e7-4452-be46-a7af14fe0d82" containerName="registry-server" Feb 20 08:33:50 crc kubenswrapper[4948]: E0220 08:33:50.999922 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26bee9f7-36e7-4452-be46-a7af14fe0d82" containerName="extract-content" Feb 20 08:33:50 crc kubenswrapper[4948]: I0220 08:33:50.999935 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="26bee9f7-36e7-4452-be46-a7af14fe0d82" containerName="extract-content" Feb 20 08:33:51 crc kubenswrapper[4948]: E0220 08:33:50.999961 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f42ddd0c-a2ec-414a-812d-f87213a3226d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.000008 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f42ddd0c-a2ec-414a-812d-f87213a3226d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.000431 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="26bee9f7-36e7-4452-be46-a7af14fe0d82" containerName="registry-server" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.000472 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f42ddd0c-a2ec-414a-812d-f87213a3226d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.001373 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.003543 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.007463 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sgx49" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.007536 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.007687 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.026551 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6"] Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.186667 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8gm4\" (UniqueName: \"kubernetes.io/projected/d38195a4-0c7b-427e-b5e2-923b24c10674-kube-api-access-p8gm4\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6\" (UID: \"d38195a4-0c7b-427e-b5e2-923b24c10674\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.186915 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d38195a4-0c7b-427e-b5e2-923b24c10674-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6\" (UID: \"d38195a4-0c7b-427e-b5e2-923b24c10674\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.187129 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d38195a4-0c7b-427e-b5e2-923b24c10674-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6\" (UID: \"d38195a4-0c7b-427e-b5e2-923b24c10674\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.289098 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d38195a4-0c7b-427e-b5e2-923b24c10674-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6\" (UID: \"d38195a4-0c7b-427e-b5e2-923b24c10674\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.289182 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d38195a4-0c7b-427e-b5e2-923b24c10674-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6\" (UID: \"d38195a4-0c7b-427e-b5e2-923b24c10674\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.289426 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8gm4\" (UniqueName: \"kubernetes.io/projected/d38195a4-0c7b-427e-b5e2-923b24c10674-kube-api-access-p8gm4\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6\" (UID: \"d38195a4-0c7b-427e-b5e2-923b24c10674\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.295810 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d38195a4-0c7b-427e-b5e2-923b24c10674-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6\" (UID: \"d38195a4-0c7b-427e-b5e2-923b24c10674\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.301085 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d38195a4-0c7b-427e-b5e2-923b24c10674-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6\" (UID: \"d38195a4-0c7b-427e-b5e2-923b24c10674\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.325307 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8gm4\" (UniqueName: \"kubernetes.io/projected/d38195a4-0c7b-427e-b5e2-923b24c10674-kube-api-access-p8gm4\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6\" (UID: \"d38195a4-0c7b-427e-b5e2-923b24c10674\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.621320 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" Feb 20 08:33:52 crc kubenswrapper[4948]: I0220 08:33:52.064255 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6"] Feb 20 08:33:52 crc kubenswrapper[4948]: I0220 08:33:52.473833 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:33:52 crc kubenswrapper[4948]: I0220 08:33:52.864779 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" event={"ID":"d38195a4-0c7b-427e-b5e2-923b24c10674","Type":"ContainerStarted","Data":"22f28812074b1280b97152f9c9a34b69bc5023788861680ff29864ecb44039df"} Feb 20 08:33:52 crc kubenswrapper[4948]: I0220 08:33:52.865354 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" event={"ID":"d38195a4-0c7b-427e-b5e2-923b24c10674","Type":"ContainerStarted","Data":"f1bf8524d66477c9c879feaa8dab8a73e46ed694dc686bb9ceaf259ba30a6e1e"} Feb 20 08:33:52 crc kubenswrapper[4948]: I0220 08:33:52.897896 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" podStartSLOduration=2.496455098 podStartE2EDuration="2.897878807s" podCreationTimestamp="2026-02-20 08:33:50 +0000 UTC" firstStartedPulling="2026-02-20 08:33:52.070076961 +0000 UTC m=+1681.044571781" lastFinishedPulling="2026-02-20 08:33:52.47150067 +0000 UTC m=+1681.445995490" observedRunningTime="2026-02-20 08:33:52.888459683 +0000 UTC m=+1681.862954503" watchObservedRunningTime="2026-02-20 08:33:52.897878807 +0000 UTC m=+1681.872373627" Feb 20 08:34:00 crc kubenswrapper[4948]: I0220 08:34:00.759316 4948 scope.go:117] "RemoveContainer" containerID="97e6bc167799c61927aab7b946388571d89bdf96214e87d2ba307142c0a90dd5" Feb 20 08:34:00 crc kubenswrapper[4948]: I0220 08:34:00.807342 4948 scope.go:117] "RemoveContainer" containerID="57a582df620782031dd517ee8549cae414c4157e56c0fc270434b874a8cbfef6" Feb 20 08:34:00 crc kubenswrapper[4948]: I0220 08:34:00.862758 4948 scope.go:117] "RemoveContainer" containerID="75747717ce5381d273c3b7d9687d0d98efbbf16a7bbea02c0a10888c6dd3a064" Feb 20 08:34:00 crc kubenswrapper[4948]: I0220 08:34:00.906947 4948 scope.go:117] "RemoveContainer" containerID="be6725bc5e2da748cc92ca6d3d54cea0585e5ca878e37159db354dc55a1eae54" Feb 20 08:34:00 crc kubenswrapper[4948]: I0220 08:34:00.955518 4948 scope.go:117] "RemoveContainer" containerID="dd9bca0f8ae0d62df16de42592e7d1d0793a43cd05a5a5d5dcaf73199e707005" Feb 20 08:34:00 crc kubenswrapper[4948]: I0220 08:34:00.997426 4948 scope.go:117] "RemoveContainer" containerID="b6240fac3873ea6c61937ec67473ce610d9e47448d45e4e229e635b6cb80cfa1" Feb 20 08:34:01 crc kubenswrapper[4948]: I0220 08:34:01.078735 4948 scope.go:117] "RemoveContainer" containerID="ca086fdde2d4c69122e9c700d4d4af3b8218c685639bf6df2d417cc08c1e51d3" Feb 20 08:34:01 crc kubenswrapper[4948]: I0220 08:34:01.112389 4948 scope.go:117] "RemoveContainer" containerID="5b0b3d5fbb40595687dd65feb8573194de8a92b25615987166c82509c29136f4" Feb 20 08:34:01 crc kubenswrapper[4948]: I0220 08:34:01.733883 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:34:01 crc kubenswrapper[4948]: E0220 08:34:01.734968 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:34:08 crc kubenswrapper[4948]: I0220 08:34:08.059163 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xt5t7"] Feb 20 08:34:08 crc kubenswrapper[4948]: I0220 08:34:08.075956 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xt5t7"] Feb 20 08:34:09 crc kubenswrapper[4948]: I0220 08:34:09.741384 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fdb2e02-8713-4205-ae31-0af33f8e4047" path="/var/lib/kubelet/pods/3fdb2e02-8713-4205-ae31-0af33f8e4047/volumes" Feb 20 08:34:12 crc kubenswrapper[4948]: I0220 08:34:12.722282 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:34:12 crc kubenswrapper[4948]: E0220 08:34:12.723339 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:34:25 crc kubenswrapper[4948]: I0220 08:34:25.723025 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:34:25 crc kubenswrapper[4948]: E0220 08:34:25.724087 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:34:29 crc kubenswrapper[4948]: I0220 08:34:29.049915 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-qnlrn"] Feb 20 08:34:29 crc kubenswrapper[4948]: I0220 08:34:29.069709 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-qnlrn"] Feb 20 08:34:29 crc kubenswrapper[4948]: I0220 08:34:29.741889 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b" path="/var/lib/kubelet/pods/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b/volumes" Feb 20 08:34:30 crc kubenswrapper[4948]: I0220 08:34:30.038088 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-jl8ck"] Feb 20 08:34:30 crc kubenswrapper[4948]: I0220 08:34:30.056727 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-jl8ck"] Feb 20 08:34:31 crc kubenswrapper[4948]: I0220 08:34:31.736682 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93f8e582-bd6a-44d4-a203-ba9950efcada" path="/var/lib/kubelet/pods/93f8e582-bd6a-44d4-a203-ba9950efcada/volumes" Feb 20 08:34:38 crc kubenswrapper[4948]: I0220 08:34:38.722914 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:34:38 crc kubenswrapper[4948]: E0220 08:34:38.724276 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:34:43 crc kubenswrapper[4948]: I0220 08:34:43.452781 4948 generic.go:334] "Generic (PLEG): container finished" podID="d38195a4-0c7b-427e-b5e2-923b24c10674" containerID="22f28812074b1280b97152f9c9a34b69bc5023788861680ff29864ecb44039df" exitCode=0 Feb 20 08:34:43 crc kubenswrapper[4948]: I0220 08:34:43.452890 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" event={"ID":"d38195a4-0c7b-427e-b5e2-923b24c10674","Type":"ContainerDied","Data":"22f28812074b1280b97152f9c9a34b69bc5023788861680ff29864ecb44039df"} Feb 20 08:34:44 crc kubenswrapper[4948]: I0220 08:34:44.976495 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.036358 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d38195a4-0c7b-427e-b5e2-923b24c10674-ssh-key-openstack-edpm-ipam\") pod \"d38195a4-0c7b-427e-b5e2-923b24c10674\" (UID: \"d38195a4-0c7b-427e-b5e2-923b24c10674\") " Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.036417 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8gm4\" (UniqueName: \"kubernetes.io/projected/d38195a4-0c7b-427e-b5e2-923b24c10674-kube-api-access-p8gm4\") pod \"d38195a4-0c7b-427e-b5e2-923b24c10674\" (UID: \"d38195a4-0c7b-427e-b5e2-923b24c10674\") " Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.036696 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d38195a4-0c7b-427e-b5e2-923b24c10674-inventory\") pod \"d38195a4-0c7b-427e-b5e2-923b24c10674\" (UID: \"d38195a4-0c7b-427e-b5e2-923b24c10674\") " Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.042616 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d38195a4-0c7b-427e-b5e2-923b24c10674-kube-api-access-p8gm4" (OuterVolumeSpecName: "kube-api-access-p8gm4") pod "d38195a4-0c7b-427e-b5e2-923b24c10674" (UID: "d38195a4-0c7b-427e-b5e2-923b24c10674"). InnerVolumeSpecName "kube-api-access-p8gm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.066162 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d38195a4-0c7b-427e-b5e2-923b24c10674-inventory" (OuterVolumeSpecName: "inventory") pod "d38195a4-0c7b-427e-b5e2-923b24c10674" (UID: "d38195a4-0c7b-427e-b5e2-923b24c10674"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.071773 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d38195a4-0c7b-427e-b5e2-923b24c10674-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "d38195a4-0c7b-427e-b5e2-923b24c10674" (UID: "d38195a4-0c7b-427e-b5e2-923b24c10674"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.138962 4948 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d38195a4-0c7b-427e-b5e2-923b24c10674-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.139011 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d38195a4-0c7b-427e-b5e2-923b24c10674-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.139023 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8gm4\" (UniqueName: \"kubernetes.io/projected/d38195a4-0c7b-427e-b5e2-923b24c10674-kube-api-access-p8gm4\") on node \"crc\" DevicePath \"\"" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.480622 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" event={"ID":"d38195a4-0c7b-427e-b5e2-923b24c10674","Type":"ContainerDied","Data":"f1bf8524d66477c9c879feaa8dab8a73e46ed694dc686bb9ceaf259ba30a6e1e"} Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.480676 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1bf8524d66477c9c879feaa8dab8a73e46ed694dc686bb9ceaf259ba30a6e1e" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.480699 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.592670 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-cqchc"] Feb 20 08:34:45 crc kubenswrapper[4948]: E0220 08:34:45.593116 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d38195a4-0c7b-427e-b5e2-923b24c10674" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.593135 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d38195a4-0c7b-427e-b5e2-923b24c10674" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.593382 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d38195a4-0c7b-427e-b5e2-923b24c10674" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.594157 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.596906 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.597350 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.597457 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.599415 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sgx49" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.603648 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-cqchc"] Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.648592 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/41856163-b621-439a-95c1-ca8d13f0a08c-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-cqchc\" (UID: \"41856163-b621-439a-95c1-ca8d13f0a08c\") " pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.648872 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/41856163-b621-439a-95c1-ca8d13f0a08c-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-cqchc\" (UID: \"41856163-b621-439a-95c1-ca8d13f0a08c\") " pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.649086 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsvbm\" (UniqueName: \"kubernetes.io/projected/41856163-b621-439a-95c1-ca8d13f0a08c-kube-api-access-fsvbm\") pod \"ssh-known-hosts-edpm-deployment-cqchc\" (UID: \"41856163-b621-439a-95c1-ca8d13f0a08c\") " pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.751327 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsvbm\" (UniqueName: \"kubernetes.io/projected/41856163-b621-439a-95c1-ca8d13f0a08c-kube-api-access-fsvbm\") pod \"ssh-known-hosts-edpm-deployment-cqchc\" (UID: \"41856163-b621-439a-95c1-ca8d13f0a08c\") " pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.753857 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/41856163-b621-439a-95c1-ca8d13f0a08c-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-cqchc\" (UID: \"41856163-b621-439a-95c1-ca8d13f0a08c\") " pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.754404 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/41856163-b621-439a-95c1-ca8d13f0a08c-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-cqchc\" (UID: \"41856163-b621-439a-95c1-ca8d13f0a08c\") " pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.757235 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/41856163-b621-439a-95c1-ca8d13f0a08c-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-cqchc\" (UID: \"41856163-b621-439a-95c1-ca8d13f0a08c\") " pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.758723 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/41856163-b621-439a-95c1-ca8d13f0a08c-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-cqchc\" (UID: \"41856163-b621-439a-95c1-ca8d13f0a08c\") " pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.774742 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsvbm\" (UniqueName: \"kubernetes.io/projected/41856163-b621-439a-95c1-ca8d13f0a08c-kube-api-access-fsvbm\") pod \"ssh-known-hosts-edpm-deployment-cqchc\" (UID: \"41856163-b621-439a-95c1-ca8d13f0a08c\") " pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.914432 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" Feb 20 08:34:46 crc kubenswrapper[4948]: I0220 08:34:46.439077 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-cqchc"] Feb 20 08:34:46 crc kubenswrapper[4948]: I0220 08:34:46.494705 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" event={"ID":"41856163-b621-439a-95c1-ca8d13f0a08c","Type":"ContainerStarted","Data":"f7c8ce188843d7fd25ac7b234d3e54139f992328b052544c97a2a433dce34ce0"} Feb 20 08:34:47 crc kubenswrapper[4948]: I0220 08:34:47.507370 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" event={"ID":"41856163-b621-439a-95c1-ca8d13f0a08c","Type":"ContainerStarted","Data":"180dc0cc05cc1856f0bc16e7267d6cbb4c78efff33f86f297dc54c1848d855ae"} Feb 20 08:34:47 crc kubenswrapper[4948]: I0220 08:34:47.533497 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" podStartSLOduration=2.004716972 podStartE2EDuration="2.533475917s" podCreationTimestamp="2026-02-20 08:34:45 +0000 UTC" firstStartedPulling="2026-02-20 08:34:46.445201645 +0000 UTC m=+1735.419696485" lastFinishedPulling="2026-02-20 08:34:46.9739606 +0000 UTC m=+1735.948455430" observedRunningTime="2026-02-20 08:34:47.529871778 +0000 UTC m=+1736.504366618" watchObservedRunningTime="2026-02-20 08:34:47.533475917 +0000 UTC m=+1736.507970747" Feb 20 08:34:49 crc kubenswrapper[4948]: I0220 08:34:49.723580 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:34:49 crc kubenswrapper[4948]: E0220 08:34:49.724346 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:34:54 crc kubenswrapper[4948]: I0220 08:34:54.574321 4948 generic.go:334] "Generic (PLEG): container finished" podID="41856163-b621-439a-95c1-ca8d13f0a08c" containerID="180dc0cc05cc1856f0bc16e7267d6cbb4c78efff33f86f297dc54c1848d855ae" exitCode=0 Feb 20 08:34:54 crc kubenswrapper[4948]: I0220 08:34:54.574741 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" event={"ID":"41856163-b621-439a-95c1-ca8d13f0a08c","Type":"ContainerDied","Data":"180dc0cc05cc1856f0bc16e7267d6cbb4c78efff33f86f297dc54c1848d855ae"} Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.039710 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.173203 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/41856163-b621-439a-95c1-ca8d13f0a08c-inventory-0\") pod \"41856163-b621-439a-95c1-ca8d13f0a08c\" (UID: \"41856163-b621-439a-95c1-ca8d13f0a08c\") " Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.173355 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/41856163-b621-439a-95c1-ca8d13f0a08c-ssh-key-openstack-edpm-ipam\") pod \"41856163-b621-439a-95c1-ca8d13f0a08c\" (UID: \"41856163-b621-439a-95c1-ca8d13f0a08c\") " Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.173505 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fsvbm\" (UniqueName: \"kubernetes.io/projected/41856163-b621-439a-95c1-ca8d13f0a08c-kube-api-access-fsvbm\") pod \"41856163-b621-439a-95c1-ca8d13f0a08c\" (UID: \"41856163-b621-439a-95c1-ca8d13f0a08c\") " Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.178720 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41856163-b621-439a-95c1-ca8d13f0a08c-kube-api-access-fsvbm" (OuterVolumeSpecName: "kube-api-access-fsvbm") pod "41856163-b621-439a-95c1-ca8d13f0a08c" (UID: "41856163-b621-439a-95c1-ca8d13f0a08c"). InnerVolumeSpecName "kube-api-access-fsvbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.200767 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41856163-b621-439a-95c1-ca8d13f0a08c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "41856163-b621-439a-95c1-ca8d13f0a08c" (UID: "41856163-b621-439a-95c1-ca8d13f0a08c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.213865 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41856163-b621-439a-95c1-ca8d13f0a08c-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "41856163-b621-439a-95c1-ca8d13f0a08c" (UID: "41856163-b621-439a-95c1-ca8d13f0a08c"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.275283 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fsvbm\" (UniqueName: \"kubernetes.io/projected/41856163-b621-439a-95c1-ca8d13f0a08c-kube-api-access-fsvbm\") on node \"crc\" DevicePath \"\"" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.275317 4948 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/41856163-b621-439a-95c1-ca8d13f0a08c-inventory-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.275326 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/41856163-b621-439a-95c1-ca8d13f0a08c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.590736 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" event={"ID":"41856163-b621-439a-95c1-ca8d13f0a08c","Type":"ContainerDied","Data":"f7c8ce188843d7fd25ac7b234d3e54139f992328b052544c97a2a433dce34ce0"} Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.590779 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7c8ce188843d7fd25ac7b234d3e54139f992328b052544c97a2a433dce34ce0" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.590808 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.683519 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf"] Feb 20 08:34:56 crc kubenswrapper[4948]: E0220 08:34:56.683952 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41856163-b621-439a-95c1-ca8d13f0a08c" containerName="ssh-known-hosts-edpm-deployment" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.684091 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="41856163-b621-439a-95c1-ca8d13f0a08c" containerName="ssh-known-hosts-edpm-deployment" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.684369 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="41856163-b621-439a-95c1-ca8d13f0a08c" containerName="ssh-known-hosts-edpm-deployment" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.685106 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.688753 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.690184 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.690732 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.690736 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sgx49" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.705110 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf"] Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.787548 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lkbkf\" (UID: \"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.787749 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lkbkf\" (UID: \"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.787850 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lxdt\" (UniqueName: \"kubernetes.io/projected/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-kube-api-access-7lxdt\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lkbkf\" (UID: \"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.890533 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lkbkf\" (UID: \"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.890897 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lkbkf\" (UID: \"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.891187 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lxdt\" (UniqueName: \"kubernetes.io/projected/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-kube-api-access-7lxdt\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lkbkf\" (UID: \"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.900496 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lkbkf\" (UID: \"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.903261 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lkbkf\" (UID: \"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.933063 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lxdt\" (UniqueName: \"kubernetes.io/projected/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-kube-api-access-7lxdt\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lkbkf\" (UID: \"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" Feb 20 08:34:57 crc kubenswrapper[4948]: I0220 08:34:57.004472 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" Feb 20 08:34:57 crc kubenswrapper[4948]: I0220 08:34:57.551737 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf"] Feb 20 08:34:57 crc kubenswrapper[4948]: I0220 08:34:57.600835 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" event={"ID":"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8","Type":"ContainerStarted","Data":"d5a7e5b1b8b0989f19fb68130988dab642729d856e78aedf79aafad894716612"} Feb 20 08:34:58 crc kubenswrapper[4948]: I0220 08:34:58.612221 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" event={"ID":"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8","Type":"ContainerStarted","Data":"40f85bc69eb4ee43a735e4ca0fe70430eaa230f800b74744b7b63478abaf2dd8"} Feb 20 08:34:58 crc kubenswrapper[4948]: I0220 08:34:58.635880 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" podStartSLOduration=2.1843637 podStartE2EDuration="2.635859696s" podCreationTimestamp="2026-02-20 08:34:56 +0000 UTC" firstStartedPulling="2026-02-20 08:34:57.546679039 +0000 UTC m=+1746.521173859" lastFinishedPulling="2026-02-20 08:34:57.998175025 +0000 UTC m=+1746.972669855" observedRunningTime="2026-02-20 08:34:58.633739683 +0000 UTC m=+1747.608234503" watchObservedRunningTime="2026-02-20 08:34:58.635859696 +0000 UTC m=+1747.610354516" Feb 20 08:35:01 crc kubenswrapper[4948]: I0220 08:35:01.399233 4948 scope.go:117] "RemoveContainer" containerID="9a9d495cbe6ddcf0246661f3edacc2b8afe21e8ef43eca82e9140a67f21f1c6d" Feb 20 08:35:01 crc kubenswrapper[4948]: I0220 08:35:01.468576 4948 scope.go:117] "RemoveContainer" containerID="67cb9d3659cf07e0efb010d75eb62a0dddad421b940140c438f0a0714676a1ed" Feb 20 08:35:01 crc kubenswrapper[4948]: I0220 08:35:01.514898 4948 scope.go:117] "RemoveContainer" containerID="53a1cca1717eaf7448d7d84b4b68c8bad0f23b2d1cc07d696f16d15b417b8356" Feb 20 08:35:01 crc kubenswrapper[4948]: I0220 08:35:01.730109 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:35:01 crc kubenswrapper[4948]: E0220 08:35:01.730440 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:35:06 crc kubenswrapper[4948]: I0220 08:35:06.696871 4948 generic.go:334] "Generic (PLEG): container finished" podID="79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8" containerID="40f85bc69eb4ee43a735e4ca0fe70430eaa230f800b74744b7b63478abaf2dd8" exitCode=0 Feb 20 08:35:06 crc kubenswrapper[4948]: I0220 08:35:06.697012 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" event={"ID":"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8","Type":"ContainerDied","Data":"40f85bc69eb4ee43a735e4ca0fe70430eaa230f800b74744b7b63478abaf2dd8"} Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.139403 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.222390 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-inventory\") pod \"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8\" (UID: \"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8\") " Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.222463 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lxdt\" (UniqueName: \"kubernetes.io/projected/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-kube-api-access-7lxdt\") pod \"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8\" (UID: \"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8\") " Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.222524 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-ssh-key-openstack-edpm-ipam\") pod \"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8\" (UID: \"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8\") " Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.227582 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-kube-api-access-7lxdt" (OuterVolumeSpecName: "kube-api-access-7lxdt") pod "79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8" (UID: "79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8"). InnerVolumeSpecName "kube-api-access-7lxdt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:35:08 crc kubenswrapper[4948]: E0220 08:35:08.252320 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-ssh-key-openstack-edpm-ipam podName:79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8 nodeName:}" failed. No retries permitted until 2026-02-20 08:35:08.752290018 +0000 UTC m=+1757.726784828 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "ssh-key-openstack-edpm-ipam" (UniqueName: "kubernetes.io/secret/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-ssh-key-openstack-edpm-ipam") pod "79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8" (UID: "79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8") : error deleting /var/lib/kubelet/pods/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8/volume-subpaths: remove /var/lib/kubelet/pods/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8/volume-subpaths: no such file or directory Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.256022 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-inventory" (OuterVolumeSpecName: "inventory") pod "79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8" (UID: "79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.328262 4948 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.328300 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lxdt\" (UniqueName: \"kubernetes.io/projected/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-kube-api-access-7lxdt\") on node \"crc\" DevicePath \"\"" Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.727054 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" event={"ID":"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8","Type":"ContainerDied","Data":"d5a7e5b1b8b0989f19fb68130988dab642729d856e78aedf79aafad894716612"} Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.727120 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5a7e5b1b8b0989f19fb68130988dab642729d856e78aedf79aafad894716612" Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.727078 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.826943 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522"] Feb 20 08:35:08 crc kubenswrapper[4948]: E0220 08:35:08.827958 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.828037 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.828483 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.829889 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.837497 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522"] Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.841440 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-ssh-key-openstack-edpm-ipam\") pod \"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8\" (UID: \"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8\") " Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.847780 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8" (UID: "79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.944489 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e3b05342-f970-4fdb-9b1c-ff521e118f96-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-ds522\" (UID: \"e3b05342-f970-4fdb-9b1c-ff521e118f96\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.944833 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3b05342-f970-4fdb-9b1c-ff521e118f96-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-ds522\" (UID: \"e3b05342-f970-4fdb-9b1c-ff521e118f96\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.945235 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjgmm\" (UniqueName: \"kubernetes.io/projected/e3b05342-f970-4fdb-9b1c-ff521e118f96-kube-api-access-pjgmm\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-ds522\" (UID: \"e3b05342-f970-4fdb-9b1c-ff521e118f96\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.945495 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:35:09 crc kubenswrapper[4948]: I0220 08:35:09.048244 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjgmm\" (UniqueName: \"kubernetes.io/projected/e3b05342-f970-4fdb-9b1c-ff521e118f96-kube-api-access-pjgmm\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-ds522\" (UID: \"e3b05342-f970-4fdb-9b1c-ff521e118f96\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" Feb 20 08:35:09 crc kubenswrapper[4948]: I0220 08:35:09.048557 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e3b05342-f970-4fdb-9b1c-ff521e118f96-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-ds522\" (UID: \"e3b05342-f970-4fdb-9b1c-ff521e118f96\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" Feb 20 08:35:09 crc kubenswrapper[4948]: I0220 08:35:09.048834 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3b05342-f970-4fdb-9b1c-ff521e118f96-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-ds522\" (UID: \"e3b05342-f970-4fdb-9b1c-ff521e118f96\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" Feb 20 08:35:09 crc kubenswrapper[4948]: I0220 08:35:09.055555 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e3b05342-f970-4fdb-9b1c-ff521e118f96-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-ds522\" (UID: \"e3b05342-f970-4fdb-9b1c-ff521e118f96\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" Feb 20 08:35:09 crc kubenswrapper[4948]: I0220 08:35:09.056928 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3b05342-f970-4fdb-9b1c-ff521e118f96-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-ds522\" (UID: \"e3b05342-f970-4fdb-9b1c-ff521e118f96\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" Feb 20 08:35:09 crc kubenswrapper[4948]: I0220 08:35:09.068825 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjgmm\" (UniqueName: \"kubernetes.io/projected/e3b05342-f970-4fdb-9b1c-ff521e118f96-kube-api-access-pjgmm\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-ds522\" (UID: \"e3b05342-f970-4fdb-9b1c-ff521e118f96\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" Feb 20 08:35:09 crc kubenswrapper[4948]: I0220 08:35:09.193990 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" Feb 20 08:35:10 crc kubenswrapper[4948]: I0220 08:35:09.998585 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522"] Feb 20 08:35:10 crc kubenswrapper[4948]: I0220 08:35:10.746270 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" event={"ID":"e3b05342-f970-4fdb-9b1c-ff521e118f96","Type":"ContainerStarted","Data":"0efd1de2d13cbcfa7bb9f687818d126c6dfe370e31b97b9d68feae9168d4d5a5"} Feb 20 08:35:10 crc kubenswrapper[4948]: I0220 08:35:10.746669 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" event={"ID":"e3b05342-f970-4fdb-9b1c-ff521e118f96","Type":"ContainerStarted","Data":"0af26b0e4338eeef5f6df928698713b0376131f10f3bb8116615033f43d3300b"} Feb 20 08:35:10 crc kubenswrapper[4948]: I0220 08:35:10.775748 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" podStartSLOduration=2.36727645 podStartE2EDuration="2.775727143s" podCreationTimestamp="2026-02-20 08:35:08 +0000 UTC" firstStartedPulling="2026-02-20 08:35:10.00318989 +0000 UTC m=+1758.977684710" lastFinishedPulling="2026-02-20 08:35:10.411640583 +0000 UTC m=+1759.386135403" observedRunningTime="2026-02-20 08:35:10.767126229 +0000 UTC m=+1759.741621049" watchObservedRunningTime="2026-02-20 08:35:10.775727143 +0000 UTC m=+1759.750221963" Feb 20 08:35:14 crc kubenswrapper[4948]: I0220 08:35:14.069050 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-mg24f"] Feb 20 08:35:14 crc kubenswrapper[4948]: I0220 08:35:14.085041 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-mg24f"] Feb 20 08:35:15 crc kubenswrapper[4948]: I0220 08:35:15.723508 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:35:15 crc kubenswrapper[4948]: E0220 08:35:15.724360 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:35:15 crc kubenswrapper[4948]: I0220 08:35:15.738295 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf549f46-7597-4399-addc-3bf3f5b63f99" path="/var/lib/kubelet/pods/cf549f46-7597-4399-addc-3bf3f5b63f99/volumes" Feb 20 08:35:20 crc kubenswrapper[4948]: I0220 08:35:20.855770 4948 generic.go:334] "Generic (PLEG): container finished" podID="e3b05342-f970-4fdb-9b1c-ff521e118f96" containerID="0efd1de2d13cbcfa7bb9f687818d126c6dfe370e31b97b9d68feae9168d4d5a5" exitCode=0 Feb 20 08:35:20 crc kubenswrapper[4948]: I0220 08:35:20.855913 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" event={"ID":"e3b05342-f970-4fdb-9b1c-ff521e118f96","Type":"ContainerDied","Data":"0efd1de2d13cbcfa7bb9f687818d126c6dfe370e31b97b9d68feae9168d4d5a5"} Feb 20 08:35:22 crc kubenswrapper[4948]: I0220 08:35:22.356957 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" Feb 20 08:35:22 crc kubenswrapper[4948]: I0220 08:35:22.542762 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3b05342-f970-4fdb-9b1c-ff521e118f96-inventory\") pod \"e3b05342-f970-4fdb-9b1c-ff521e118f96\" (UID: \"e3b05342-f970-4fdb-9b1c-ff521e118f96\") " Feb 20 08:35:22 crc kubenswrapper[4948]: I0220 08:35:22.542918 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e3b05342-f970-4fdb-9b1c-ff521e118f96-ssh-key-openstack-edpm-ipam\") pod \"e3b05342-f970-4fdb-9b1c-ff521e118f96\" (UID: \"e3b05342-f970-4fdb-9b1c-ff521e118f96\") " Feb 20 08:35:22 crc kubenswrapper[4948]: I0220 08:35:22.543034 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjgmm\" (UniqueName: \"kubernetes.io/projected/e3b05342-f970-4fdb-9b1c-ff521e118f96-kube-api-access-pjgmm\") pod \"e3b05342-f970-4fdb-9b1c-ff521e118f96\" (UID: \"e3b05342-f970-4fdb-9b1c-ff521e118f96\") " Feb 20 08:35:22 crc kubenswrapper[4948]: I0220 08:35:22.552209 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3b05342-f970-4fdb-9b1c-ff521e118f96-kube-api-access-pjgmm" (OuterVolumeSpecName: "kube-api-access-pjgmm") pod "e3b05342-f970-4fdb-9b1c-ff521e118f96" (UID: "e3b05342-f970-4fdb-9b1c-ff521e118f96"). InnerVolumeSpecName "kube-api-access-pjgmm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:35:22 crc kubenswrapper[4948]: I0220 08:35:22.583175 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3b05342-f970-4fdb-9b1c-ff521e118f96-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "e3b05342-f970-4fdb-9b1c-ff521e118f96" (UID: "e3b05342-f970-4fdb-9b1c-ff521e118f96"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:35:22 crc kubenswrapper[4948]: I0220 08:35:22.617134 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3b05342-f970-4fdb-9b1c-ff521e118f96-inventory" (OuterVolumeSpecName: "inventory") pod "e3b05342-f970-4fdb-9b1c-ff521e118f96" (UID: "e3b05342-f970-4fdb-9b1c-ff521e118f96"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:35:22 crc kubenswrapper[4948]: I0220 08:35:22.648328 4948 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3b05342-f970-4fdb-9b1c-ff521e118f96-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 08:35:22 crc kubenswrapper[4948]: I0220 08:35:22.648362 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e3b05342-f970-4fdb-9b1c-ff521e118f96-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:35:22 crc kubenswrapper[4948]: I0220 08:35:22.648372 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjgmm\" (UniqueName: \"kubernetes.io/projected/e3b05342-f970-4fdb-9b1c-ff521e118f96-kube-api-access-pjgmm\") on node \"crc\" DevicePath \"\"" Feb 20 08:35:22 crc kubenswrapper[4948]: I0220 08:35:22.888012 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" Feb 20 08:35:22 crc kubenswrapper[4948]: I0220 08:35:22.891217 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" event={"ID":"e3b05342-f970-4fdb-9b1c-ff521e118f96","Type":"ContainerDied","Data":"0af26b0e4338eeef5f6df928698713b0376131f10f3bb8116615033f43d3300b"} Feb 20 08:35:22 crc kubenswrapper[4948]: I0220 08:35:22.891272 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0af26b0e4338eeef5f6df928698713b0376131f10f3bb8116615033f43d3300b" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.009182 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw"] Feb 20 08:35:23 crc kubenswrapper[4948]: E0220 08:35:23.009736 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3b05342-f970-4fdb-9b1c-ff521e118f96" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.009762 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3b05342-f970-4fdb-9b1c-ff521e118f96" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.010105 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3b05342-f970-4fdb-9b1c-ff521e118f96" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.011603 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.014121 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.014417 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.015056 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.015105 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.015227 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.015301 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.015600 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.015613 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sgx49" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.030742 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw"] Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.158485 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.158532 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.158611 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.158650 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.158673 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.158696 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.158750 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.158776 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lwbl\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-kube-api-access-4lwbl\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.158792 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.158814 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.159085 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.159181 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.159265 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.159320 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.262264 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.262424 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.262513 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.262559 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.262699 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.262754 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lwbl\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-kube-api-access-4lwbl\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.262790 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.262863 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.263772 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.263841 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.263947 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.264005 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.264067 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.264114 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.268434 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.268587 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.269040 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.269220 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.269557 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.270087 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.270948 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.271598 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.272536 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.274251 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.281379 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.281639 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.282115 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.286244 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lwbl\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-kube-api-access-4lwbl\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.337805 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: W0220 08:35:23.888786 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc517042a_5354_4a8d_b7ff_c9aafa263b6c.slice/crio-835aa7fbf4e7707cff80ac88d34e7f17f33d8271cc0cd95b2c1162f2db4176aa WatchSource:0}: Error finding container 835aa7fbf4e7707cff80ac88d34e7f17f33d8271cc0cd95b2c1162f2db4176aa: Status 404 returned error can't find the container with id 835aa7fbf4e7707cff80ac88d34e7f17f33d8271cc0cd95b2c1162f2db4176aa Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.904477 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw"] Feb 20 08:35:24 crc kubenswrapper[4948]: I0220 08:35:24.907421 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" event={"ID":"c517042a-5354-4a8d-b7ff-c9aafa263b6c","Type":"ContainerStarted","Data":"cc1c5248f2eeca932bf8566e96f4d010e0d2b251e1788d5c187c35a67338cca0"} Feb 20 08:35:24 crc kubenswrapper[4948]: I0220 08:35:24.907773 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" event={"ID":"c517042a-5354-4a8d-b7ff-c9aafa263b6c","Type":"ContainerStarted","Data":"835aa7fbf4e7707cff80ac88d34e7f17f33d8271cc0cd95b2c1162f2db4176aa"} Feb 20 08:35:24 crc kubenswrapper[4948]: I0220 08:35:24.928051 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" podStartSLOduration=2.531815314 podStartE2EDuration="2.928024752s" podCreationTimestamp="2026-02-20 08:35:22 +0000 UTC" firstStartedPulling="2026-02-20 08:35:23.892144914 +0000 UTC m=+1772.866639744" lastFinishedPulling="2026-02-20 08:35:24.288354322 +0000 UTC m=+1773.262849182" observedRunningTime="2026-02-20 08:35:24.92633649 +0000 UTC m=+1773.900831330" watchObservedRunningTime="2026-02-20 08:35:24.928024752 +0000 UTC m=+1773.902519612" Feb 20 08:35:26 crc kubenswrapper[4948]: I0220 08:35:26.723834 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:35:26 crc kubenswrapper[4948]: E0220 08:35:26.724594 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:35:41 crc kubenswrapper[4948]: I0220 08:35:41.732664 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:35:42 crc kubenswrapper[4948]: I0220 08:35:42.115016 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"df133c52e6c7c766068e5890945bf88c2349bad99b1086eddf879822786cef6f"} Feb 20 08:36:01 crc kubenswrapper[4948]: I0220 08:36:01.613602 4948 scope.go:117] "RemoveContainer" containerID="d68a0fb50da62e47bddf140b8e1628a7abf7e9ac68f7c4c4efaa6f9ccf62f179" Feb 20 08:36:02 crc kubenswrapper[4948]: I0220 08:36:02.318394 4948 generic.go:334] "Generic (PLEG): container finished" podID="c517042a-5354-4a8d-b7ff-c9aafa263b6c" containerID="cc1c5248f2eeca932bf8566e96f4d010e0d2b251e1788d5c187c35a67338cca0" exitCode=0 Feb 20 08:36:02 crc kubenswrapper[4948]: I0220 08:36:02.318532 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" event={"ID":"c517042a-5354-4a8d-b7ff-c9aafa263b6c","Type":"ContainerDied","Data":"cc1c5248f2eeca932bf8566e96f4d010e0d2b251e1788d5c187c35a67338cca0"} Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.880728 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.945020 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.945075 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-telemetry-combined-ca-bundle\") pod \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.945102 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-repo-setup-combined-ca-bundle\") pod \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.945126 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-ovn-combined-ca-bundle\") pod \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.945159 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.945198 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-libvirt-combined-ca-bundle\") pod \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.945236 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-nova-combined-ca-bundle\") pod \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.945258 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-ovn-default-certs-0\") pod \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.946363 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.946493 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-inventory\") pod \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.946534 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-ssh-key-openstack-edpm-ipam\") pod \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.946571 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lwbl\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-kube-api-access-4lwbl\") pod \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.946599 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-bootstrap-combined-ca-bundle\") pod \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.946626 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-neutron-metadata-combined-ca-bundle\") pod \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.952675 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "c517042a-5354-4a8d-b7ff-c9aafa263b6c" (UID: "c517042a-5354-4a8d-b7ff-c9aafa263b6c"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.953936 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "c517042a-5354-4a8d-b7ff-c9aafa263b6c" (UID: "c517042a-5354-4a8d-b7ff-c9aafa263b6c"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.954147 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "c517042a-5354-4a8d-b7ff-c9aafa263b6c" (UID: "c517042a-5354-4a8d-b7ff-c9aafa263b6c"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.954771 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "c517042a-5354-4a8d-b7ff-c9aafa263b6c" (UID: "c517042a-5354-4a8d-b7ff-c9aafa263b6c"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.955617 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "c517042a-5354-4a8d-b7ff-c9aafa263b6c" (UID: "c517042a-5354-4a8d-b7ff-c9aafa263b6c"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.955713 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "c517042a-5354-4a8d-b7ff-c9aafa263b6c" (UID: "c517042a-5354-4a8d-b7ff-c9aafa263b6c"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.957586 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "c517042a-5354-4a8d-b7ff-c9aafa263b6c" (UID: "c517042a-5354-4a8d-b7ff-c9aafa263b6c"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.957633 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "c517042a-5354-4a8d-b7ff-c9aafa263b6c" (UID: "c517042a-5354-4a8d-b7ff-c9aafa263b6c"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.960061 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "c517042a-5354-4a8d-b7ff-c9aafa263b6c" (UID: "c517042a-5354-4a8d-b7ff-c9aafa263b6c"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.960290 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "c517042a-5354-4a8d-b7ff-c9aafa263b6c" (UID: "c517042a-5354-4a8d-b7ff-c9aafa263b6c"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.961217 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-kube-api-access-4lwbl" (OuterVolumeSpecName: "kube-api-access-4lwbl") pod "c517042a-5354-4a8d-b7ff-c9aafa263b6c" (UID: "c517042a-5354-4a8d-b7ff-c9aafa263b6c"). InnerVolumeSpecName "kube-api-access-4lwbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.965323 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "c517042a-5354-4a8d-b7ff-c9aafa263b6c" (UID: "c517042a-5354-4a8d-b7ff-c9aafa263b6c"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.994516 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-inventory" (OuterVolumeSpecName: "inventory") pod "c517042a-5354-4a8d-b7ff-c9aafa263b6c" (UID: "c517042a-5354-4a8d-b7ff-c9aafa263b6c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.009925 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "c517042a-5354-4a8d-b7ff-c9aafa263b6c" (UID: "c517042a-5354-4a8d-b7ff-c9aafa263b6c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.049021 4948 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.049228 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.049349 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lwbl\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-kube-api-access-4lwbl\") on node \"crc\" DevicePath \"\"" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.049487 4948 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.049600 4948 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.049715 4948 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.049900 4948 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.050107 4948 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.050248 4948 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.050376 4948 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.050492 4948 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.050604 4948 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.050715 4948 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.050824 4948 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.343515 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" event={"ID":"c517042a-5354-4a8d-b7ff-c9aafa263b6c","Type":"ContainerDied","Data":"835aa7fbf4e7707cff80ac88d34e7f17f33d8271cc0cd95b2c1162f2db4176aa"} Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.343933 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="835aa7fbf4e7707cff80ac88d34e7f17f33d8271cc0cd95b2c1162f2db4176aa" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.343577 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.468319 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp"] Feb 20 08:36:04 crc kubenswrapper[4948]: E0220 08:36:04.468785 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c517042a-5354-4a8d-b7ff-c9aafa263b6c" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.468812 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="c517042a-5354-4a8d-b7ff-c9aafa263b6c" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.469070 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="c517042a-5354-4a8d-b7ff-c9aafa263b6c" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.469834 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.473298 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sgx49" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.473414 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.473304 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.474397 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.476194 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.485045 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp"] Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.561148 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4b2vp\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.561209 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4b2vp\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.561263 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mw76\" (UniqueName: \"kubernetes.io/projected/b19ca7cb-01d2-4965-90f6-5d50211870cc-kube-api-access-7mw76\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4b2vp\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.561281 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4b2vp\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.561324 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b19ca7cb-01d2-4965-90f6-5d50211870cc-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4b2vp\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.662511 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4b2vp\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.662599 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4b2vp\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.662673 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mw76\" (UniqueName: \"kubernetes.io/projected/b19ca7cb-01d2-4965-90f6-5d50211870cc-kube-api-access-7mw76\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4b2vp\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.662697 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4b2vp\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.662754 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b19ca7cb-01d2-4965-90f6-5d50211870cc-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4b2vp\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.664006 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b19ca7cb-01d2-4965-90f6-5d50211870cc-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4b2vp\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.666577 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4b2vp\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.667152 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4b2vp\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.667615 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4b2vp\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.684347 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mw76\" (UniqueName: \"kubernetes.io/projected/b19ca7cb-01d2-4965-90f6-5d50211870cc-kube-api-access-7mw76\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4b2vp\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.789493 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:36:05 crc kubenswrapper[4948]: I0220 08:36:05.353340 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp"] Feb 20 08:36:06 crc kubenswrapper[4948]: I0220 08:36:06.375141 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" event={"ID":"b19ca7cb-01d2-4965-90f6-5d50211870cc","Type":"ContainerStarted","Data":"3579159916fda6b8e313e027e891a641d962daeb27067a651a9d7a959ebf4296"} Feb 20 08:36:06 crc kubenswrapper[4948]: I0220 08:36:06.375715 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" event={"ID":"b19ca7cb-01d2-4965-90f6-5d50211870cc","Type":"ContainerStarted","Data":"03b19942c3e65950a34fa710ca9783f811998ce8fb63c17f12d02bef5e823c2b"} Feb 20 08:36:06 crc kubenswrapper[4948]: I0220 08:36:06.398863 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" podStartSLOduration=1.837710511 podStartE2EDuration="2.398845589s" podCreationTimestamp="2026-02-20 08:36:04 +0000 UTC" firstStartedPulling="2026-02-20 08:36:05.378967588 +0000 UTC m=+1814.353462408" lastFinishedPulling="2026-02-20 08:36:05.940102666 +0000 UTC m=+1814.914597486" observedRunningTime="2026-02-20 08:36:06.390092872 +0000 UTC m=+1815.364587732" watchObservedRunningTime="2026-02-20 08:36:06.398845589 +0000 UTC m=+1815.373340409" Feb 20 08:37:10 crc kubenswrapper[4948]: I0220 08:37:10.016466 4948 generic.go:334] "Generic (PLEG): container finished" podID="b19ca7cb-01d2-4965-90f6-5d50211870cc" containerID="3579159916fda6b8e313e027e891a641d962daeb27067a651a9d7a959ebf4296" exitCode=0 Feb 20 08:37:10 crc kubenswrapper[4948]: I0220 08:37:10.016577 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" event={"ID":"b19ca7cb-01d2-4965-90f6-5d50211870cc","Type":"ContainerDied","Data":"3579159916fda6b8e313e027e891a641d962daeb27067a651a9d7a959ebf4296"} Feb 20 08:37:11 crc kubenswrapper[4948]: I0220 08:37:11.541597 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:37:11 crc kubenswrapper[4948]: I0220 08:37:11.717020 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-inventory\") pod \"b19ca7cb-01d2-4965-90f6-5d50211870cc\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " Feb 20 08:37:11 crc kubenswrapper[4948]: I0220 08:37:11.717324 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b19ca7cb-01d2-4965-90f6-5d50211870cc-ovncontroller-config-0\") pod \"b19ca7cb-01d2-4965-90f6-5d50211870cc\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " Feb 20 08:37:11 crc kubenswrapper[4948]: I0220 08:37:11.717388 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-ovn-combined-ca-bundle\") pod \"b19ca7cb-01d2-4965-90f6-5d50211870cc\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " Feb 20 08:37:11 crc kubenswrapper[4948]: I0220 08:37:11.717496 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mw76\" (UniqueName: \"kubernetes.io/projected/b19ca7cb-01d2-4965-90f6-5d50211870cc-kube-api-access-7mw76\") pod \"b19ca7cb-01d2-4965-90f6-5d50211870cc\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " Feb 20 08:37:11 crc kubenswrapper[4948]: I0220 08:37:11.717662 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-ssh-key-openstack-edpm-ipam\") pod \"b19ca7cb-01d2-4965-90f6-5d50211870cc\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " Feb 20 08:37:11 crc kubenswrapper[4948]: I0220 08:37:11.736120 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b19ca7cb-01d2-4965-90f6-5d50211870cc-kube-api-access-7mw76" (OuterVolumeSpecName: "kube-api-access-7mw76") pod "b19ca7cb-01d2-4965-90f6-5d50211870cc" (UID: "b19ca7cb-01d2-4965-90f6-5d50211870cc"). InnerVolumeSpecName "kube-api-access-7mw76". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:37:11 crc kubenswrapper[4948]: I0220 08:37:11.737501 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "b19ca7cb-01d2-4965-90f6-5d50211870cc" (UID: "b19ca7cb-01d2-4965-90f6-5d50211870cc"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:37:11 crc kubenswrapper[4948]: I0220 08:37:11.747660 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-inventory" (OuterVolumeSpecName: "inventory") pod "b19ca7cb-01d2-4965-90f6-5d50211870cc" (UID: "b19ca7cb-01d2-4965-90f6-5d50211870cc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:37:11 crc kubenswrapper[4948]: I0220 08:37:11.764818 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "b19ca7cb-01d2-4965-90f6-5d50211870cc" (UID: "b19ca7cb-01d2-4965-90f6-5d50211870cc"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:37:11 crc kubenswrapper[4948]: I0220 08:37:11.779251 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b19ca7cb-01d2-4965-90f6-5d50211870cc-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "b19ca7cb-01d2-4965-90f6-5d50211870cc" (UID: "b19ca7cb-01d2-4965-90f6-5d50211870cc"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:37:11 crc kubenswrapper[4948]: I0220 08:37:11.821009 4948 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b19ca7cb-01d2-4965-90f6-5d50211870cc-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:37:11 crc kubenswrapper[4948]: I0220 08:37:11.821080 4948 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:37:11 crc kubenswrapper[4948]: I0220 08:37:11.821109 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mw76\" (UniqueName: \"kubernetes.io/projected/b19ca7cb-01d2-4965-90f6-5d50211870cc-kube-api-access-7mw76\") on node \"crc\" DevicePath \"\"" Feb 20 08:37:11 crc kubenswrapper[4948]: I0220 08:37:11.821137 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:37:11 crc kubenswrapper[4948]: I0220 08:37:11.821165 4948 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.044536 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" event={"ID":"b19ca7cb-01d2-4965-90f6-5d50211870cc","Type":"ContainerDied","Data":"03b19942c3e65950a34fa710ca9783f811998ce8fb63c17f12d02bef5e823c2b"} Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.044598 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03b19942c3e65950a34fa710ca9783f811998ce8fb63c17f12d02bef5e823c2b" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.044662 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.209884 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr"] Feb 20 08:37:12 crc kubenswrapper[4948]: E0220 08:37:12.210695 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b19ca7cb-01d2-4965-90f6-5d50211870cc" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.210712 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b19ca7cb-01d2-4965-90f6-5d50211870cc" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.210960 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="b19ca7cb-01d2-4965-90f6-5d50211870cc" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.211719 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.217011 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sgx49" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.217182 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.217280 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.217385 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.217494 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.217263 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.219864 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr"] Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.333156 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.333418 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.333583 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.333674 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.333783 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.334163 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cp9df\" (UniqueName: \"kubernetes.io/projected/fd280342-4276-4eee-a763-876e31be28c5-kube-api-access-cp9df\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.435571 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.435824 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.435933 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.436028 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cp9df\" (UniqueName: \"kubernetes.io/projected/fd280342-4276-4eee-a763-876e31be28c5-kube-api-access-cp9df\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.436124 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.436209 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.440773 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.441047 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.441467 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.442286 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.451481 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.454255 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cp9df\" (UniqueName: \"kubernetes.io/projected/fd280342-4276-4eee-a763-876e31be28c5-kube-api-access-cp9df\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.539767 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:13 crc kubenswrapper[4948]: I0220 08:37:13.151203 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr"] Feb 20 08:37:14 crc kubenswrapper[4948]: I0220 08:37:14.070101 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" event={"ID":"fd280342-4276-4eee-a763-876e31be28c5","Type":"ContainerStarted","Data":"da36425e07d1062948867ebf1c1a2889de547fdeccbc610aebafa16fe2b84d8c"} Feb 20 08:37:14 crc kubenswrapper[4948]: I0220 08:37:14.070506 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" event={"ID":"fd280342-4276-4eee-a763-876e31be28c5","Type":"ContainerStarted","Data":"59f5aaf9278d5e594e4ceb19412a5a664c76f9cc8655925b2ee3a4a3ea09e8b3"} Feb 20 08:37:14 crc kubenswrapper[4948]: I0220 08:37:14.098397 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" podStartSLOduration=1.63944544 podStartE2EDuration="2.098372453s" podCreationTimestamp="2026-02-20 08:37:12 +0000 UTC" firstStartedPulling="2026-02-20 08:37:13.171435483 +0000 UTC m=+1882.145930313" lastFinishedPulling="2026-02-20 08:37:13.630362476 +0000 UTC m=+1882.604857326" observedRunningTime="2026-02-20 08:37:14.097283966 +0000 UTC m=+1883.071778846" watchObservedRunningTime="2026-02-20 08:37:14.098372453 +0000 UTC m=+1883.072867303" Feb 20 08:37:30 crc kubenswrapper[4948]: I0220 08:37:30.751678 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-cqm4v"] Feb 20 08:37:30 crc kubenswrapper[4948]: I0220 08:37:30.755783 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cqm4v" Feb 20 08:37:30 crc kubenswrapper[4948]: I0220 08:37:30.766219 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cqm4v"] Feb 20 08:37:30 crc kubenswrapper[4948]: I0220 08:37:30.849100 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35e908da-6c2c-4e75-a567-7ad1b95dacf4-utilities\") pod \"redhat-marketplace-cqm4v\" (UID: \"35e908da-6c2c-4e75-a567-7ad1b95dacf4\") " pod="openshift-marketplace/redhat-marketplace-cqm4v" Feb 20 08:37:30 crc kubenswrapper[4948]: I0220 08:37:30.849189 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjrcl\" (UniqueName: \"kubernetes.io/projected/35e908da-6c2c-4e75-a567-7ad1b95dacf4-kube-api-access-hjrcl\") pod \"redhat-marketplace-cqm4v\" (UID: \"35e908da-6c2c-4e75-a567-7ad1b95dacf4\") " pod="openshift-marketplace/redhat-marketplace-cqm4v" Feb 20 08:37:30 crc kubenswrapper[4948]: I0220 08:37:30.849279 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35e908da-6c2c-4e75-a567-7ad1b95dacf4-catalog-content\") pod \"redhat-marketplace-cqm4v\" (UID: \"35e908da-6c2c-4e75-a567-7ad1b95dacf4\") " pod="openshift-marketplace/redhat-marketplace-cqm4v" Feb 20 08:37:30 crc kubenswrapper[4948]: I0220 08:37:30.951464 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjrcl\" (UniqueName: \"kubernetes.io/projected/35e908da-6c2c-4e75-a567-7ad1b95dacf4-kube-api-access-hjrcl\") pod \"redhat-marketplace-cqm4v\" (UID: \"35e908da-6c2c-4e75-a567-7ad1b95dacf4\") " pod="openshift-marketplace/redhat-marketplace-cqm4v" Feb 20 08:37:30 crc kubenswrapper[4948]: I0220 08:37:30.951899 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35e908da-6c2c-4e75-a567-7ad1b95dacf4-catalog-content\") pod \"redhat-marketplace-cqm4v\" (UID: \"35e908da-6c2c-4e75-a567-7ad1b95dacf4\") " pod="openshift-marketplace/redhat-marketplace-cqm4v" Feb 20 08:37:30 crc kubenswrapper[4948]: I0220 08:37:30.952190 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35e908da-6c2c-4e75-a567-7ad1b95dacf4-utilities\") pod \"redhat-marketplace-cqm4v\" (UID: \"35e908da-6c2c-4e75-a567-7ad1b95dacf4\") " pod="openshift-marketplace/redhat-marketplace-cqm4v" Feb 20 08:37:30 crc kubenswrapper[4948]: I0220 08:37:30.952784 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35e908da-6c2c-4e75-a567-7ad1b95dacf4-catalog-content\") pod \"redhat-marketplace-cqm4v\" (UID: \"35e908da-6c2c-4e75-a567-7ad1b95dacf4\") " pod="openshift-marketplace/redhat-marketplace-cqm4v" Feb 20 08:37:30 crc kubenswrapper[4948]: I0220 08:37:30.952812 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35e908da-6c2c-4e75-a567-7ad1b95dacf4-utilities\") pod \"redhat-marketplace-cqm4v\" (UID: \"35e908da-6c2c-4e75-a567-7ad1b95dacf4\") " pod="openshift-marketplace/redhat-marketplace-cqm4v" Feb 20 08:37:30 crc kubenswrapper[4948]: I0220 08:37:30.982421 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjrcl\" (UniqueName: \"kubernetes.io/projected/35e908da-6c2c-4e75-a567-7ad1b95dacf4-kube-api-access-hjrcl\") pod \"redhat-marketplace-cqm4v\" (UID: \"35e908da-6c2c-4e75-a567-7ad1b95dacf4\") " pod="openshift-marketplace/redhat-marketplace-cqm4v" Feb 20 08:37:31 crc kubenswrapper[4948]: I0220 08:37:31.093032 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cqm4v" Feb 20 08:37:31 crc kubenswrapper[4948]: I0220 08:37:31.645465 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cqm4v"] Feb 20 08:37:32 crc kubenswrapper[4948]: I0220 08:37:32.249110 4948 generic.go:334] "Generic (PLEG): container finished" podID="35e908da-6c2c-4e75-a567-7ad1b95dacf4" containerID="cc4ee197b9a059416805589950a2af69d52f9384246abc136a26690a03566cc8" exitCode=0 Feb 20 08:37:32 crc kubenswrapper[4948]: I0220 08:37:32.249227 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cqm4v" event={"ID":"35e908da-6c2c-4e75-a567-7ad1b95dacf4","Type":"ContainerDied","Data":"cc4ee197b9a059416805589950a2af69d52f9384246abc136a26690a03566cc8"} Feb 20 08:37:32 crc kubenswrapper[4948]: I0220 08:37:32.249500 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cqm4v" event={"ID":"35e908da-6c2c-4e75-a567-7ad1b95dacf4","Type":"ContainerStarted","Data":"8bcf259501cac28be63c9a620abde433c308b13a601573640c689320a460160d"} Feb 20 08:37:33 crc kubenswrapper[4948]: I0220 08:37:33.268661 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cqm4v" event={"ID":"35e908da-6c2c-4e75-a567-7ad1b95dacf4","Type":"ContainerStarted","Data":"35e1c15903ef72b4f5767d562be34dd2b8ca0c9c4040a73fcefd14e2c16aa84b"} Feb 20 08:37:34 crc kubenswrapper[4948]: I0220 08:37:34.284416 4948 generic.go:334] "Generic (PLEG): container finished" podID="35e908da-6c2c-4e75-a567-7ad1b95dacf4" containerID="35e1c15903ef72b4f5767d562be34dd2b8ca0c9c4040a73fcefd14e2c16aa84b" exitCode=0 Feb 20 08:37:34 crc kubenswrapper[4948]: I0220 08:37:34.284523 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cqm4v" event={"ID":"35e908da-6c2c-4e75-a567-7ad1b95dacf4","Type":"ContainerDied","Data":"35e1c15903ef72b4f5767d562be34dd2b8ca0c9c4040a73fcefd14e2c16aa84b"} Feb 20 08:37:35 crc kubenswrapper[4948]: I0220 08:37:35.296597 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cqm4v" event={"ID":"35e908da-6c2c-4e75-a567-7ad1b95dacf4","Type":"ContainerStarted","Data":"b75e9f035613e5a0bca9527f0d1752421f2a649641c9c5f2a53156d6a7934952"} Feb 20 08:37:35 crc kubenswrapper[4948]: I0220 08:37:35.330640 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-cqm4v" podStartSLOduration=2.8606195420000002 podStartE2EDuration="5.330620569s" podCreationTimestamp="2026-02-20 08:37:30 +0000 UTC" firstStartedPulling="2026-02-20 08:37:32.251540499 +0000 UTC m=+1901.226035399" lastFinishedPulling="2026-02-20 08:37:34.721541576 +0000 UTC m=+1903.696036426" observedRunningTime="2026-02-20 08:37:35.322468478 +0000 UTC m=+1904.296963318" watchObservedRunningTime="2026-02-20 08:37:35.330620569 +0000 UTC m=+1904.305115399" Feb 20 08:37:41 crc kubenswrapper[4948]: I0220 08:37:41.093245 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-cqm4v" Feb 20 08:37:41 crc kubenswrapper[4948]: I0220 08:37:41.094164 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-cqm4v" Feb 20 08:37:41 crc kubenswrapper[4948]: I0220 08:37:41.170861 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-cqm4v" Feb 20 08:37:41 crc kubenswrapper[4948]: I0220 08:37:41.422498 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-cqm4v" Feb 20 08:37:41 crc kubenswrapper[4948]: I0220 08:37:41.475644 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cqm4v"] Feb 20 08:37:43 crc kubenswrapper[4948]: I0220 08:37:43.375810 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-cqm4v" podUID="35e908da-6c2c-4e75-a567-7ad1b95dacf4" containerName="registry-server" containerID="cri-o://b75e9f035613e5a0bca9527f0d1752421f2a649641c9c5f2a53156d6a7934952" gracePeriod=2 Feb 20 08:37:43 crc kubenswrapper[4948]: I0220 08:37:43.929272 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cqm4v" Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.091951 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35e908da-6c2c-4e75-a567-7ad1b95dacf4-utilities\") pod \"35e908da-6c2c-4e75-a567-7ad1b95dacf4\" (UID: \"35e908da-6c2c-4e75-a567-7ad1b95dacf4\") " Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.092109 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjrcl\" (UniqueName: \"kubernetes.io/projected/35e908da-6c2c-4e75-a567-7ad1b95dacf4-kube-api-access-hjrcl\") pod \"35e908da-6c2c-4e75-a567-7ad1b95dacf4\" (UID: \"35e908da-6c2c-4e75-a567-7ad1b95dacf4\") " Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.092471 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35e908da-6c2c-4e75-a567-7ad1b95dacf4-catalog-content\") pod \"35e908da-6c2c-4e75-a567-7ad1b95dacf4\" (UID: \"35e908da-6c2c-4e75-a567-7ad1b95dacf4\") " Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.093428 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35e908da-6c2c-4e75-a567-7ad1b95dacf4-utilities" (OuterVolumeSpecName: "utilities") pod "35e908da-6c2c-4e75-a567-7ad1b95dacf4" (UID: "35e908da-6c2c-4e75-a567-7ad1b95dacf4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.102472 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35e908da-6c2c-4e75-a567-7ad1b95dacf4-kube-api-access-hjrcl" (OuterVolumeSpecName: "kube-api-access-hjrcl") pod "35e908da-6c2c-4e75-a567-7ad1b95dacf4" (UID: "35e908da-6c2c-4e75-a567-7ad1b95dacf4"). InnerVolumeSpecName "kube-api-access-hjrcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.123166 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35e908da-6c2c-4e75-a567-7ad1b95dacf4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "35e908da-6c2c-4e75-a567-7ad1b95dacf4" (UID: "35e908da-6c2c-4e75-a567-7ad1b95dacf4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.197191 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35e908da-6c2c-4e75-a567-7ad1b95dacf4-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.197231 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35e908da-6c2c-4e75-a567-7ad1b95dacf4-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.197244 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjrcl\" (UniqueName: \"kubernetes.io/projected/35e908da-6c2c-4e75-a567-7ad1b95dacf4-kube-api-access-hjrcl\") on node \"crc\" DevicePath \"\"" Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.394851 4948 generic.go:334] "Generic (PLEG): container finished" podID="35e908da-6c2c-4e75-a567-7ad1b95dacf4" containerID="b75e9f035613e5a0bca9527f0d1752421f2a649641c9c5f2a53156d6a7934952" exitCode=0 Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.394966 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cqm4v" Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.395046 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cqm4v" event={"ID":"35e908da-6c2c-4e75-a567-7ad1b95dacf4","Type":"ContainerDied","Data":"b75e9f035613e5a0bca9527f0d1752421f2a649641c9c5f2a53156d6a7934952"} Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.396525 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cqm4v" event={"ID":"35e908da-6c2c-4e75-a567-7ad1b95dacf4","Type":"ContainerDied","Data":"8bcf259501cac28be63c9a620abde433c308b13a601573640c689320a460160d"} Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.396567 4948 scope.go:117] "RemoveContainer" containerID="b75e9f035613e5a0bca9527f0d1752421f2a649641c9c5f2a53156d6a7934952" Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.425878 4948 scope.go:117] "RemoveContainer" containerID="35e1c15903ef72b4f5767d562be34dd2b8ca0c9c4040a73fcefd14e2c16aa84b" Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.460580 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cqm4v"] Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.471788 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-cqm4v"] Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.472262 4948 scope.go:117] "RemoveContainer" containerID="cc4ee197b9a059416805589950a2af69d52f9384246abc136a26690a03566cc8" Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.516323 4948 scope.go:117] "RemoveContainer" containerID="b75e9f035613e5a0bca9527f0d1752421f2a649641c9c5f2a53156d6a7934952" Feb 20 08:37:44 crc kubenswrapper[4948]: E0220 08:37:44.520741 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b75e9f035613e5a0bca9527f0d1752421f2a649641c9c5f2a53156d6a7934952\": container with ID starting with b75e9f035613e5a0bca9527f0d1752421f2a649641c9c5f2a53156d6a7934952 not found: ID does not exist" containerID="b75e9f035613e5a0bca9527f0d1752421f2a649641c9c5f2a53156d6a7934952" Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.520791 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b75e9f035613e5a0bca9527f0d1752421f2a649641c9c5f2a53156d6a7934952"} err="failed to get container status \"b75e9f035613e5a0bca9527f0d1752421f2a649641c9c5f2a53156d6a7934952\": rpc error: code = NotFound desc = could not find container \"b75e9f035613e5a0bca9527f0d1752421f2a649641c9c5f2a53156d6a7934952\": container with ID starting with b75e9f035613e5a0bca9527f0d1752421f2a649641c9c5f2a53156d6a7934952 not found: ID does not exist" Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.520817 4948 scope.go:117] "RemoveContainer" containerID="35e1c15903ef72b4f5767d562be34dd2b8ca0c9c4040a73fcefd14e2c16aa84b" Feb 20 08:37:44 crc kubenswrapper[4948]: E0220 08:37:44.521306 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35e1c15903ef72b4f5767d562be34dd2b8ca0c9c4040a73fcefd14e2c16aa84b\": container with ID starting with 35e1c15903ef72b4f5767d562be34dd2b8ca0c9c4040a73fcefd14e2c16aa84b not found: ID does not exist" containerID="35e1c15903ef72b4f5767d562be34dd2b8ca0c9c4040a73fcefd14e2c16aa84b" Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.521355 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35e1c15903ef72b4f5767d562be34dd2b8ca0c9c4040a73fcefd14e2c16aa84b"} err="failed to get container status \"35e1c15903ef72b4f5767d562be34dd2b8ca0c9c4040a73fcefd14e2c16aa84b\": rpc error: code = NotFound desc = could not find container \"35e1c15903ef72b4f5767d562be34dd2b8ca0c9c4040a73fcefd14e2c16aa84b\": container with ID starting with 35e1c15903ef72b4f5767d562be34dd2b8ca0c9c4040a73fcefd14e2c16aa84b not found: ID does not exist" Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.521398 4948 scope.go:117] "RemoveContainer" containerID="cc4ee197b9a059416805589950a2af69d52f9384246abc136a26690a03566cc8" Feb 20 08:37:44 crc kubenswrapper[4948]: E0220 08:37:44.521811 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc4ee197b9a059416805589950a2af69d52f9384246abc136a26690a03566cc8\": container with ID starting with cc4ee197b9a059416805589950a2af69d52f9384246abc136a26690a03566cc8 not found: ID does not exist" containerID="cc4ee197b9a059416805589950a2af69d52f9384246abc136a26690a03566cc8" Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.521856 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc4ee197b9a059416805589950a2af69d52f9384246abc136a26690a03566cc8"} err="failed to get container status \"cc4ee197b9a059416805589950a2af69d52f9384246abc136a26690a03566cc8\": rpc error: code = NotFound desc = could not find container \"cc4ee197b9a059416805589950a2af69d52f9384246abc136a26690a03566cc8\": container with ID starting with cc4ee197b9a059416805589950a2af69d52f9384246abc136a26690a03566cc8 not found: ID does not exist" Feb 20 08:37:45 crc kubenswrapper[4948]: I0220 08:37:45.733993 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35e908da-6c2c-4e75-a567-7ad1b95dacf4" path="/var/lib/kubelet/pods/35e908da-6c2c-4e75-a567-7ad1b95dacf4/volumes" Feb 20 08:38:02 crc kubenswrapper[4948]: I0220 08:38:02.596384 4948 generic.go:334] "Generic (PLEG): container finished" podID="fd280342-4276-4eee-a763-876e31be28c5" containerID="da36425e07d1062948867ebf1c1a2889de547fdeccbc610aebafa16fe2b84d8c" exitCode=0 Feb 20 08:38:02 crc kubenswrapper[4948]: I0220 08:38:02.596468 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" event={"ID":"fd280342-4276-4eee-a763-876e31be28c5","Type":"ContainerDied","Data":"da36425e07d1062948867ebf1c1a2889de547fdeccbc610aebafa16fe2b84d8c"} Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.210869 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.377784 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-inventory\") pod \"fd280342-4276-4eee-a763-876e31be28c5\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.378288 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-neutron-metadata-combined-ca-bundle\") pod \"fd280342-4276-4eee-a763-876e31be28c5\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.378339 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cp9df\" (UniqueName: \"kubernetes.io/projected/fd280342-4276-4eee-a763-876e31be28c5-kube-api-access-cp9df\") pod \"fd280342-4276-4eee-a763-876e31be28c5\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.378566 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-ssh-key-openstack-edpm-ipam\") pod \"fd280342-4276-4eee-a763-876e31be28c5\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.378688 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-neutron-ovn-metadata-agent-neutron-config-0\") pod \"fd280342-4276-4eee-a763-876e31be28c5\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.378794 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-nova-metadata-neutron-config-0\") pod \"fd280342-4276-4eee-a763-876e31be28c5\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.385747 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd280342-4276-4eee-a763-876e31be28c5-kube-api-access-cp9df" (OuterVolumeSpecName: "kube-api-access-cp9df") pod "fd280342-4276-4eee-a763-876e31be28c5" (UID: "fd280342-4276-4eee-a763-876e31be28c5"). InnerVolumeSpecName "kube-api-access-cp9df". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.385807 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "fd280342-4276-4eee-a763-876e31be28c5" (UID: "fd280342-4276-4eee-a763-876e31be28c5"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.418115 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-inventory" (OuterVolumeSpecName: "inventory") pod "fd280342-4276-4eee-a763-876e31be28c5" (UID: "fd280342-4276-4eee-a763-876e31be28c5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.419516 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "fd280342-4276-4eee-a763-876e31be28c5" (UID: "fd280342-4276-4eee-a763-876e31be28c5"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.429558 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "fd280342-4276-4eee-a763-876e31be28c5" (UID: "fd280342-4276-4eee-a763-876e31be28c5"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.441902 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "fd280342-4276-4eee-a763-876e31be28c5" (UID: "fd280342-4276-4eee-a763-876e31be28c5"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.482175 4948 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.482215 4948 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.482276 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cp9df\" (UniqueName: \"kubernetes.io/projected/fd280342-4276-4eee-a763-876e31be28c5-kube-api-access-cp9df\") on node \"crc\" DevicePath \"\"" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.482288 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.482323 4948 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.482337 4948 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.623260 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" event={"ID":"fd280342-4276-4eee-a763-876e31be28c5","Type":"ContainerDied","Data":"59f5aaf9278d5e594e4ceb19412a5a664c76f9cc8655925b2ee3a4a3ea09e8b3"} Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.623329 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59f5aaf9278d5e594e4ceb19412a5a664c76f9cc8655925b2ee3a4a3ea09e8b3" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.623334 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.727015 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs"] Feb 20 08:38:04 crc kubenswrapper[4948]: E0220 08:38:04.727695 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35e908da-6c2c-4e75-a567-7ad1b95dacf4" containerName="extract-content" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.727788 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="35e908da-6c2c-4e75-a567-7ad1b95dacf4" containerName="extract-content" Feb 20 08:38:04 crc kubenswrapper[4948]: E0220 08:38:04.727876 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35e908da-6c2c-4e75-a567-7ad1b95dacf4" containerName="registry-server" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.727954 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="35e908da-6c2c-4e75-a567-7ad1b95dacf4" containerName="registry-server" Feb 20 08:38:04 crc kubenswrapper[4948]: E0220 08:38:04.728068 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd280342-4276-4eee-a763-876e31be28c5" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.728140 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd280342-4276-4eee-a763-876e31be28c5" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 20 08:38:04 crc kubenswrapper[4948]: E0220 08:38:04.728222 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35e908da-6c2c-4e75-a567-7ad1b95dacf4" containerName="extract-utilities" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.728289 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="35e908da-6c2c-4e75-a567-7ad1b95dacf4" containerName="extract-utilities" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.728607 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd280342-4276-4eee-a763-876e31be28c5" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.728706 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="35e908da-6c2c-4e75-a567-7ad1b95dacf4" containerName="registry-server" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.729491 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.732561 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sgx49" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.733237 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.734147 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.741834 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs"] Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.742295 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.745555 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.788714 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.788810 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.788940 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.789075 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8gnt\" (UniqueName: \"kubernetes.io/projected/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-kube-api-access-x8gnt\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.789201 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.890186 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.890277 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8gnt\" (UniqueName: \"kubernetes.io/projected/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-kube-api-access-x8gnt\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.890347 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.890609 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.890664 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.895904 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.896095 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.897248 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.897382 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.925366 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8gnt\" (UniqueName: \"kubernetes.io/projected/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-kube-api-access-x8gnt\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:38:05 crc kubenswrapper[4948]: I0220 08:38:05.048938 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:38:05 crc kubenswrapper[4948]: I0220 08:38:05.595036 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs"] Feb 20 08:38:05 crc kubenswrapper[4948]: W0220 08:38:05.603183 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e619eac_dfdc_4c8c_aa51_fc08ba58b49a.slice/crio-bed584d1cffc5c44267aa8da79eb83c0547d67f904d60a696d0a5f0983625d4a WatchSource:0}: Error finding container bed584d1cffc5c44267aa8da79eb83c0547d67f904d60a696d0a5f0983625d4a: Status 404 returned error can't find the container with id bed584d1cffc5c44267aa8da79eb83c0547d67f904d60a696d0a5f0983625d4a Feb 20 08:38:05 crc kubenswrapper[4948]: I0220 08:38:05.607247 4948 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 08:38:05 crc kubenswrapper[4948]: I0220 08:38:05.632956 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" event={"ID":"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a","Type":"ContainerStarted","Data":"bed584d1cffc5c44267aa8da79eb83c0547d67f904d60a696d0a5f0983625d4a"} Feb 20 08:38:06 crc kubenswrapper[4948]: I0220 08:38:06.648881 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" event={"ID":"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a","Type":"ContainerStarted","Data":"e26f1bc45952ca3a3414ed06ca781d747c45268c9f970dd0c34eb0f719572756"} Feb 20 08:38:06 crc kubenswrapper[4948]: I0220 08:38:06.688495 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" podStartSLOduration=2.252412711 podStartE2EDuration="2.688463518s" podCreationTimestamp="2026-02-20 08:38:04 +0000 UTC" firstStartedPulling="2026-02-20 08:38:05.606999149 +0000 UTC m=+1934.581493969" lastFinishedPulling="2026-02-20 08:38:06.043049946 +0000 UTC m=+1935.017544776" observedRunningTime="2026-02-20 08:38:06.675775225 +0000 UTC m=+1935.650270075" watchObservedRunningTime="2026-02-20 08:38:06.688463518 +0000 UTC m=+1935.662958378" Feb 20 08:38:08 crc kubenswrapper[4948]: I0220 08:38:08.025442 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:38:08 crc kubenswrapper[4948]: I0220 08:38:08.026627 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:38:38 crc kubenswrapper[4948]: I0220 08:38:38.025295 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:38:38 crc kubenswrapper[4948]: I0220 08:38:38.025837 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:39:08 crc kubenswrapper[4948]: I0220 08:39:08.024572 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:39:08 crc kubenswrapper[4948]: I0220 08:39:08.025402 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:39:08 crc kubenswrapper[4948]: I0220 08:39:08.025520 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:39:08 crc kubenswrapper[4948]: I0220 08:39:08.026750 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"df133c52e6c7c766068e5890945bf88c2349bad99b1086eddf879822786cef6f"} pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 08:39:08 crc kubenswrapper[4948]: I0220 08:39:08.026853 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" containerID="cri-o://df133c52e6c7c766068e5890945bf88c2349bad99b1086eddf879822786cef6f" gracePeriod=600 Feb 20 08:39:08 crc kubenswrapper[4948]: I0220 08:39:08.300900 4948 generic.go:334] "Generic (PLEG): container finished" podID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerID="df133c52e6c7c766068e5890945bf88c2349bad99b1086eddf879822786cef6f" exitCode=0 Feb 20 08:39:08 crc kubenswrapper[4948]: I0220 08:39:08.301012 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerDied","Data":"df133c52e6c7c766068e5890945bf88c2349bad99b1086eddf879822786cef6f"} Feb 20 08:39:08 crc kubenswrapper[4948]: I0220 08:39:08.301515 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:39:09 crc kubenswrapper[4948]: I0220 08:39:09.317471 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707"} Feb 20 08:41:08 crc kubenswrapper[4948]: I0220 08:41:08.024850 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:41:08 crc kubenswrapper[4948]: I0220 08:41:08.025488 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:41:38 crc kubenswrapper[4948]: I0220 08:41:38.025140 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:41:38 crc kubenswrapper[4948]: I0220 08:41:38.025726 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:41:56 crc kubenswrapper[4948]: I0220 08:41:56.097322 4948 generic.go:334] "Generic (PLEG): container finished" podID="6e619eac-dfdc-4c8c-aa51-fc08ba58b49a" containerID="e26f1bc45952ca3a3414ed06ca781d747c45268c9f970dd0c34eb0f719572756" exitCode=0 Feb 20 08:41:56 crc kubenswrapper[4948]: I0220 08:41:56.097478 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" event={"ID":"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a","Type":"ContainerDied","Data":"e26f1bc45952ca3a3414ed06ca781d747c45268c9f970dd0c34eb0f719572756"} Feb 20 08:41:57 crc kubenswrapper[4948]: I0220 08:41:57.636697 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:41:57 crc kubenswrapper[4948]: I0220 08:41:57.790206 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-libvirt-secret-0\") pod \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " Feb 20 08:41:57 crc kubenswrapper[4948]: I0220 08:41:57.790406 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-ssh-key-openstack-edpm-ipam\") pod \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " Feb 20 08:41:57 crc kubenswrapper[4948]: I0220 08:41:57.790543 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-libvirt-combined-ca-bundle\") pod \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " Feb 20 08:41:57 crc kubenswrapper[4948]: I0220 08:41:57.790582 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8gnt\" (UniqueName: \"kubernetes.io/projected/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-kube-api-access-x8gnt\") pod \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " Feb 20 08:41:57 crc kubenswrapper[4948]: I0220 08:41:57.790624 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-inventory\") pod \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " Feb 20 08:41:57 crc kubenswrapper[4948]: I0220 08:41:57.798178 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "6e619eac-dfdc-4c8c-aa51-fc08ba58b49a" (UID: "6e619eac-dfdc-4c8c-aa51-fc08ba58b49a"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:41:57 crc kubenswrapper[4948]: I0220 08:41:57.798252 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-kube-api-access-x8gnt" (OuterVolumeSpecName: "kube-api-access-x8gnt") pod "6e619eac-dfdc-4c8c-aa51-fc08ba58b49a" (UID: "6e619eac-dfdc-4c8c-aa51-fc08ba58b49a"). InnerVolumeSpecName "kube-api-access-x8gnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:41:57 crc kubenswrapper[4948]: I0220 08:41:57.824493 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-inventory" (OuterVolumeSpecName: "inventory") pod "6e619eac-dfdc-4c8c-aa51-fc08ba58b49a" (UID: "6e619eac-dfdc-4c8c-aa51-fc08ba58b49a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:41:57 crc kubenswrapper[4948]: I0220 08:41:57.830669 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "6e619eac-dfdc-4c8c-aa51-fc08ba58b49a" (UID: "6e619eac-dfdc-4c8c-aa51-fc08ba58b49a"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:41:57 crc kubenswrapper[4948]: I0220 08:41:57.831233 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "6e619eac-dfdc-4c8c-aa51-fc08ba58b49a" (UID: "6e619eac-dfdc-4c8c-aa51-fc08ba58b49a"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:41:57 crc kubenswrapper[4948]: I0220 08:41:57.896162 4948 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:41:57 crc kubenswrapper[4948]: I0220 08:41:57.896240 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:41:57 crc kubenswrapper[4948]: I0220 08:41:57.896265 4948 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:41:57 crc kubenswrapper[4948]: I0220 08:41:57.896283 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8gnt\" (UniqueName: \"kubernetes.io/projected/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-kube-api-access-x8gnt\") on node \"crc\" DevicePath \"\"" Feb 20 08:41:57 crc kubenswrapper[4948]: I0220 08:41:57.896333 4948 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.121847 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" event={"ID":"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a","Type":"ContainerDied","Data":"bed584d1cffc5c44267aa8da79eb83c0547d67f904d60a696d0a5f0983625d4a"} Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.121948 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bed584d1cffc5c44267aa8da79eb83c0547d67f904d60a696d0a5f0983625d4a" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.122147 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.260311 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr"] Feb 20 08:41:58 crc kubenswrapper[4948]: E0220 08:41:58.261243 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e619eac-dfdc-4c8c-aa51-fc08ba58b49a" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.261262 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e619eac-dfdc-4c8c-aa51-fc08ba58b49a" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.261551 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e619eac-dfdc-4c8c-aa51-fc08ba58b49a" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.262318 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.265195 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.266067 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.266297 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.267080 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sgx49" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.267293 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.267429 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.267832 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.277524 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr"] Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.404755 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.405089 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.405216 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.405361 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.405464 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.405607 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ljgt\" (UniqueName: \"kubernetes.io/projected/6005ded0-9e21-4908-bbf0-33c710ba4341-kube-api-access-8ljgt\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.405739 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.405902 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.406081 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.406181 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.406266 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.508465 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.508638 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.508681 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.508760 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.508815 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.508853 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ljgt\" (UniqueName: \"kubernetes.io/projected/6005ded0-9e21-4908-bbf0-33c710ba4341-kube-api-access-8ljgt\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.508926 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.509051 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.509097 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.509137 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.509177 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.510208 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.512362 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.512945 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.513774 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.514102 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.515300 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.515452 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.516161 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.530702 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.533374 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.537705 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ljgt\" (UniqueName: \"kubernetes.io/projected/6005ded0-9e21-4908-bbf0-33c710ba4341-kube-api-access-8ljgt\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.589637 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:59 crc kubenswrapper[4948]: I0220 08:41:59.196170 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr"] Feb 20 08:42:00 crc kubenswrapper[4948]: I0220 08:42:00.154928 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" event={"ID":"6005ded0-9e21-4908-bbf0-33c710ba4341","Type":"ContainerStarted","Data":"2f2f83e7b2bae4f94ae51a61eb3c4783c5606e273956b4e39c8926e563dd497a"} Feb 20 08:42:00 crc kubenswrapper[4948]: I0220 08:42:00.155507 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" event={"ID":"6005ded0-9e21-4908-bbf0-33c710ba4341","Type":"ContainerStarted","Data":"f847f6d88e05df089793a5fb35fc239a9b46a7a0da87f3039554ac3950729267"} Feb 20 08:42:00 crc kubenswrapper[4948]: I0220 08:42:00.187950 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" podStartSLOduration=1.693300743 podStartE2EDuration="2.187931383s" podCreationTimestamp="2026-02-20 08:41:58 +0000 UTC" firstStartedPulling="2026-02-20 08:41:59.206956337 +0000 UTC m=+2168.181451157" lastFinishedPulling="2026-02-20 08:41:59.701586947 +0000 UTC m=+2168.676081797" observedRunningTime="2026-02-20 08:42:00.177130476 +0000 UTC m=+2169.151625316" watchObservedRunningTime="2026-02-20 08:42:00.187931383 +0000 UTC m=+2169.162426203" Feb 20 08:42:08 crc kubenswrapper[4948]: I0220 08:42:08.024731 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:42:08 crc kubenswrapper[4948]: I0220 08:42:08.025352 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:42:08 crc kubenswrapper[4948]: I0220 08:42:08.025405 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:42:08 crc kubenswrapper[4948]: I0220 08:42:08.026300 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707"} pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 08:42:08 crc kubenswrapper[4948]: I0220 08:42:08.026367 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" containerID="cri-o://4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" gracePeriod=600 Feb 20 08:42:08 crc kubenswrapper[4948]: E0220 08:42:08.157321 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:42:08 crc kubenswrapper[4948]: I0220 08:42:08.268392 4948 generic.go:334] "Generic (PLEG): container finished" podID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" exitCode=0 Feb 20 08:42:08 crc kubenswrapper[4948]: I0220 08:42:08.268468 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerDied","Data":"4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707"} Feb 20 08:42:08 crc kubenswrapper[4948]: I0220 08:42:08.268508 4948 scope.go:117] "RemoveContainer" containerID="df133c52e6c7c766068e5890945bf88c2349bad99b1086eddf879822786cef6f" Feb 20 08:42:08 crc kubenswrapper[4948]: I0220 08:42:08.271408 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:42:08 crc kubenswrapper[4948]: E0220 08:42:08.272195 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:42:22 crc kubenswrapper[4948]: I0220 08:42:22.725185 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:42:22 crc kubenswrapper[4948]: E0220 08:42:22.726115 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:42:37 crc kubenswrapper[4948]: I0220 08:42:37.729556 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:42:37 crc kubenswrapper[4948]: E0220 08:42:37.730225 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:42:42 crc kubenswrapper[4948]: I0220 08:42:42.802818 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-88wr7"] Feb 20 08:42:42 crc kubenswrapper[4948]: I0220 08:42:42.808551 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-88wr7" Feb 20 08:42:42 crc kubenswrapper[4948]: I0220 08:42:42.822364 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-88wr7"] Feb 20 08:42:42 crc kubenswrapper[4948]: I0220 08:42:42.917512 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-catalog-content\") pod \"redhat-operators-88wr7\" (UID: \"4bc1b6a7-dfaf-4c10-88be-f482a5c58667\") " pod="openshift-marketplace/redhat-operators-88wr7" Feb 20 08:42:42 crc kubenswrapper[4948]: I0220 08:42:42.917581 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gg2dj\" (UniqueName: \"kubernetes.io/projected/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-kube-api-access-gg2dj\") pod \"redhat-operators-88wr7\" (UID: \"4bc1b6a7-dfaf-4c10-88be-f482a5c58667\") " pod="openshift-marketplace/redhat-operators-88wr7" Feb 20 08:42:42 crc kubenswrapper[4948]: I0220 08:42:42.917633 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-utilities\") pod \"redhat-operators-88wr7\" (UID: \"4bc1b6a7-dfaf-4c10-88be-f482a5c58667\") " pod="openshift-marketplace/redhat-operators-88wr7" Feb 20 08:42:43 crc kubenswrapper[4948]: I0220 08:42:43.019538 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gg2dj\" (UniqueName: \"kubernetes.io/projected/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-kube-api-access-gg2dj\") pod \"redhat-operators-88wr7\" (UID: \"4bc1b6a7-dfaf-4c10-88be-f482a5c58667\") " pod="openshift-marketplace/redhat-operators-88wr7" Feb 20 08:42:43 crc kubenswrapper[4948]: I0220 08:42:43.019599 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-utilities\") pod \"redhat-operators-88wr7\" (UID: \"4bc1b6a7-dfaf-4c10-88be-f482a5c58667\") " pod="openshift-marketplace/redhat-operators-88wr7" Feb 20 08:42:43 crc kubenswrapper[4948]: I0220 08:42:43.019737 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-catalog-content\") pod \"redhat-operators-88wr7\" (UID: \"4bc1b6a7-dfaf-4c10-88be-f482a5c58667\") " pod="openshift-marketplace/redhat-operators-88wr7" Feb 20 08:42:43 crc kubenswrapper[4948]: I0220 08:42:43.020268 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-catalog-content\") pod \"redhat-operators-88wr7\" (UID: \"4bc1b6a7-dfaf-4c10-88be-f482a5c58667\") " pod="openshift-marketplace/redhat-operators-88wr7" Feb 20 08:42:43 crc kubenswrapper[4948]: I0220 08:42:43.020524 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-utilities\") pod \"redhat-operators-88wr7\" (UID: \"4bc1b6a7-dfaf-4c10-88be-f482a5c58667\") " pod="openshift-marketplace/redhat-operators-88wr7" Feb 20 08:42:43 crc kubenswrapper[4948]: I0220 08:42:43.046744 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gg2dj\" (UniqueName: \"kubernetes.io/projected/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-kube-api-access-gg2dj\") pod \"redhat-operators-88wr7\" (UID: \"4bc1b6a7-dfaf-4c10-88be-f482a5c58667\") " pod="openshift-marketplace/redhat-operators-88wr7" Feb 20 08:42:43 crc kubenswrapper[4948]: I0220 08:42:43.149288 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-88wr7" Feb 20 08:42:43 crc kubenswrapper[4948]: I0220 08:42:43.640378 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-88wr7"] Feb 20 08:42:44 crc kubenswrapper[4948]: I0220 08:42:44.633670 4948 generic.go:334] "Generic (PLEG): container finished" podID="4bc1b6a7-dfaf-4c10-88be-f482a5c58667" containerID="020c49ea97fd0d2abaf92d2a4ecacff3dea8fe68fd14d2c0bda270118de765ca" exitCode=0 Feb 20 08:42:44 crc kubenswrapper[4948]: I0220 08:42:44.633809 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-88wr7" event={"ID":"4bc1b6a7-dfaf-4c10-88be-f482a5c58667","Type":"ContainerDied","Data":"020c49ea97fd0d2abaf92d2a4ecacff3dea8fe68fd14d2c0bda270118de765ca"} Feb 20 08:42:44 crc kubenswrapper[4948]: I0220 08:42:44.634011 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-88wr7" event={"ID":"4bc1b6a7-dfaf-4c10-88be-f482a5c58667","Type":"ContainerStarted","Data":"5c209d2da2d8783cbae32471f9af6f43d7c3ec2277c0cd928735be812a4502bc"} Feb 20 08:42:45 crc kubenswrapper[4948]: I0220 08:42:45.649810 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-88wr7" event={"ID":"4bc1b6a7-dfaf-4c10-88be-f482a5c58667","Type":"ContainerStarted","Data":"7ca67f0db7f8dd9db858f07ad294b0aa2364e71f346b50828413c9852296cb73"} Feb 20 08:42:46 crc kubenswrapper[4948]: I0220 08:42:46.667738 4948 generic.go:334] "Generic (PLEG): container finished" podID="4bc1b6a7-dfaf-4c10-88be-f482a5c58667" containerID="7ca67f0db7f8dd9db858f07ad294b0aa2364e71f346b50828413c9852296cb73" exitCode=0 Feb 20 08:42:46 crc kubenswrapper[4948]: I0220 08:42:46.667832 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-88wr7" event={"ID":"4bc1b6a7-dfaf-4c10-88be-f482a5c58667","Type":"ContainerDied","Data":"7ca67f0db7f8dd9db858f07ad294b0aa2364e71f346b50828413c9852296cb73"} Feb 20 08:42:47 crc kubenswrapper[4948]: I0220 08:42:47.680801 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-88wr7" event={"ID":"4bc1b6a7-dfaf-4c10-88be-f482a5c58667","Type":"ContainerStarted","Data":"f9d71de10f48d2736cee59456b183da30aed2ce7e43fe91b4241385713ea2e1d"} Feb 20 08:42:51 crc kubenswrapper[4948]: I0220 08:42:51.735670 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:42:51 crc kubenswrapper[4948]: E0220 08:42:51.736961 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:42:53 crc kubenswrapper[4948]: I0220 08:42:53.149532 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-88wr7" Feb 20 08:42:53 crc kubenswrapper[4948]: I0220 08:42:53.149576 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-88wr7" Feb 20 08:42:54 crc kubenswrapper[4948]: I0220 08:42:54.220637 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-88wr7" podUID="4bc1b6a7-dfaf-4c10-88be-f482a5c58667" containerName="registry-server" probeResult="failure" output=< Feb 20 08:42:54 crc kubenswrapper[4948]: timeout: failed to connect service ":50051" within 1s Feb 20 08:42:54 crc kubenswrapper[4948]: > Feb 20 08:43:03 crc kubenswrapper[4948]: I0220 08:43:03.241590 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-88wr7" Feb 20 08:43:03 crc kubenswrapper[4948]: I0220 08:43:03.268725 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-88wr7" podStartSLOduration=18.807062757 podStartE2EDuration="21.268694457s" podCreationTimestamp="2026-02-20 08:42:42 +0000 UTC" firstStartedPulling="2026-02-20 08:42:44.636705196 +0000 UTC m=+2213.611200056" lastFinishedPulling="2026-02-20 08:42:47.098336896 +0000 UTC m=+2216.072831756" observedRunningTime="2026-02-20 08:42:47.708618342 +0000 UTC m=+2216.683113172" watchObservedRunningTime="2026-02-20 08:43:03.268694457 +0000 UTC m=+2232.243189317" Feb 20 08:43:03 crc kubenswrapper[4948]: I0220 08:43:03.309125 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-88wr7" Feb 20 08:43:03 crc kubenswrapper[4948]: I0220 08:43:03.484861 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-88wr7"] Feb 20 08:43:04 crc kubenswrapper[4948]: I0220 08:43:04.869772 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-88wr7" podUID="4bc1b6a7-dfaf-4c10-88be-f482a5c58667" containerName="registry-server" containerID="cri-o://f9d71de10f48d2736cee59456b183da30aed2ce7e43fe91b4241385713ea2e1d" gracePeriod=2 Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.298297 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-88wr7" Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.450722 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-catalog-content\") pod \"4bc1b6a7-dfaf-4c10-88be-f482a5c58667\" (UID: \"4bc1b6a7-dfaf-4c10-88be-f482a5c58667\") " Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.450829 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-utilities\") pod \"4bc1b6a7-dfaf-4c10-88be-f482a5c58667\" (UID: \"4bc1b6a7-dfaf-4c10-88be-f482a5c58667\") " Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.451157 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gg2dj\" (UniqueName: \"kubernetes.io/projected/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-kube-api-access-gg2dj\") pod \"4bc1b6a7-dfaf-4c10-88be-f482a5c58667\" (UID: \"4bc1b6a7-dfaf-4c10-88be-f482a5c58667\") " Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.451919 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-utilities" (OuterVolumeSpecName: "utilities") pod "4bc1b6a7-dfaf-4c10-88be-f482a5c58667" (UID: "4bc1b6a7-dfaf-4c10-88be-f482a5c58667"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.457195 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-kube-api-access-gg2dj" (OuterVolumeSpecName: "kube-api-access-gg2dj") pod "4bc1b6a7-dfaf-4c10-88be-f482a5c58667" (UID: "4bc1b6a7-dfaf-4c10-88be-f482a5c58667"). InnerVolumeSpecName "kube-api-access-gg2dj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.553788 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.553833 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gg2dj\" (UniqueName: \"kubernetes.io/projected/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-kube-api-access-gg2dj\") on node \"crc\" DevicePath \"\"" Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.567752 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4bc1b6a7-dfaf-4c10-88be-f482a5c58667" (UID: "4bc1b6a7-dfaf-4c10-88be-f482a5c58667"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.655291 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.881219 4948 generic.go:334] "Generic (PLEG): container finished" podID="4bc1b6a7-dfaf-4c10-88be-f482a5c58667" containerID="f9d71de10f48d2736cee59456b183da30aed2ce7e43fe91b4241385713ea2e1d" exitCode=0 Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.881280 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-88wr7" event={"ID":"4bc1b6a7-dfaf-4c10-88be-f482a5c58667","Type":"ContainerDied","Data":"f9d71de10f48d2736cee59456b183da30aed2ce7e43fe91b4241385713ea2e1d"} Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.881309 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-88wr7" event={"ID":"4bc1b6a7-dfaf-4c10-88be-f482a5c58667","Type":"ContainerDied","Data":"5c209d2da2d8783cbae32471f9af6f43d7c3ec2277c0cd928735be812a4502bc"} Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.881348 4948 scope.go:117] "RemoveContainer" containerID="f9d71de10f48d2736cee59456b183da30aed2ce7e43fe91b4241385713ea2e1d" Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.881481 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-88wr7" Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.911538 4948 scope.go:117] "RemoveContainer" containerID="7ca67f0db7f8dd9db858f07ad294b0aa2364e71f346b50828413c9852296cb73" Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.913069 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-88wr7"] Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.920568 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-88wr7"] Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.938540 4948 scope.go:117] "RemoveContainer" containerID="020c49ea97fd0d2abaf92d2a4ecacff3dea8fe68fd14d2c0bda270118de765ca" Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.984565 4948 scope.go:117] "RemoveContainer" containerID="f9d71de10f48d2736cee59456b183da30aed2ce7e43fe91b4241385713ea2e1d" Feb 20 08:43:05 crc kubenswrapper[4948]: E0220 08:43:05.985222 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9d71de10f48d2736cee59456b183da30aed2ce7e43fe91b4241385713ea2e1d\": container with ID starting with f9d71de10f48d2736cee59456b183da30aed2ce7e43fe91b4241385713ea2e1d not found: ID does not exist" containerID="f9d71de10f48d2736cee59456b183da30aed2ce7e43fe91b4241385713ea2e1d" Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.985290 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9d71de10f48d2736cee59456b183da30aed2ce7e43fe91b4241385713ea2e1d"} err="failed to get container status \"f9d71de10f48d2736cee59456b183da30aed2ce7e43fe91b4241385713ea2e1d\": rpc error: code = NotFound desc = could not find container \"f9d71de10f48d2736cee59456b183da30aed2ce7e43fe91b4241385713ea2e1d\": container with ID starting with f9d71de10f48d2736cee59456b183da30aed2ce7e43fe91b4241385713ea2e1d not found: ID does not exist" Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.985322 4948 scope.go:117] "RemoveContainer" containerID="7ca67f0db7f8dd9db858f07ad294b0aa2364e71f346b50828413c9852296cb73" Feb 20 08:43:05 crc kubenswrapper[4948]: E0220 08:43:05.985771 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ca67f0db7f8dd9db858f07ad294b0aa2364e71f346b50828413c9852296cb73\": container with ID starting with 7ca67f0db7f8dd9db858f07ad294b0aa2364e71f346b50828413c9852296cb73 not found: ID does not exist" containerID="7ca67f0db7f8dd9db858f07ad294b0aa2364e71f346b50828413c9852296cb73" Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.985804 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ca67f0db7f8dd9db858f07ad294b0aa2364e71f346b50828413c9852296cb73"} err="failed to get container status \"7ca67f0db7f8dd9db858f07ad294b0aa2364e71f346b50828413c9852296cb73\": rpc error: code = NotFound desc = could not find container \"7ca67f0db7f8dd9db858f07ad294b0aa2364e71f346b50828413c9852296cb73\": container with ID starting with 7ca67f0db7f8dd9db858f07ad294b0aa2364e71f346b50828413c9852296cb73 not found: ID does not exist" Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.985828 4948 scope.go:117] "RemoveContainer" containerID="020c49ea97fd0d2abaf92d2a4ecacff3dea8fe68fd14d2c0bda270118de765ca" Feb 20 08:43:05 crc kubenswrapper[4948]: E0220 08:43:05.986088 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"020c49ea97fd0d2abaf92d2a4ecacff3dea8fe68fd14d2c0bda270118de765ca\": container with ID starting with 020c49ea97fd0d2abaf92d2a4ecacff3dea8fe68fd14d2c0bda270118de765ca not found: ID does not exist" containerID="020c49ea97fd0d2abaf92d2a4ecacff3dea8fe68fd14d2c0bda270118de765ca" Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.986121 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"020c49ea97fd0d2abaf92d2a4ecacff3dea8fe68fd14d2c0bda270118de765ca"} err="failed to get container status \"020c49ea97fd0d2abaf92d2a4ecacff3dea8fe68fd14d2c0bda270118de765ca\": rpc error: code = NotFound desc = could not find container \"020c49ea97fd0d2abaf92d2a4ecacff3dea8fe68fd14d2c0bda270118de765ca\": container with ID starting with 020c49ea97fd0d2abaf92d2a4ecacff3dea8fe68fd14d2c0bda270118de765ca not found: ID does not exist" Feb 20 08:43:06 crc kubenswrapper[4948]: I0220 08:43:06.722424 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:43:06 crc kubenswrapper[4948]: E0220 08:43:06.722758 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:43:07 crc kubenswrapper[4948]: I0220 08:43:07.737619 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bc1b6a7-dfaf-4c10-88be-f482a5c58667" path="/var/lib/kubelet/pods/4bc1b6a7-dfaf-4c10-88be-f482a5c58667/volumes" Feb 20 08:43:17 crc kubenswrapper[4948]: I0220 08:43:17.723619 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:43:17 crc kubenswrapper[4948]: E0220 08:43:17.725308 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:43:29 crc kubenswrapper[4948]: I0220 08:43:29.722902 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:43:29 crc kubenswrapper[4948]: E0220 08:43:29.723587 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:43:44 crc kubenswrapper[4948]: I0220 08:43:44.722986 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:43:44 crc kubenswrapper[4948]: E0220 08:43:44.723609 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:43:56 crc kubenswrapper[4948]: I0220 08:43:56.723511 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:43:56 crc kubenswrapper[4948]: E0220 08:43:56.724840 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.015366 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vqtpd"] Feb 20 08:43:59 crc kubenswrapper[4948]: E0220 08:43:59.016362 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bc1b6a7-dfaf-4c10-88be-f482a5c58667" containerName="registry-server" Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.016384 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bc1b6a7-dfaf-4c10-88be-f482a5c58667" containerName="registry-server" Feb 20 08:43:59 crc kubenswrapper[4948]: E0220 08:43:59.016432 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bc1b6a7-dfaf-4c10-88be-f482a5c58667" containerName="extract-content" Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.016444 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bc1b6a7-dfaf-4c10-88be-f482a5c58667" containerName="extract-content" Feb 20 08:43:59 crc kubenswrapper[4948]: E0220 08:43:59.016465 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bc1b6a7-dfaf-4c10-88be-f482a5c58667" containerName="extract-utilities" Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.016477 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bc1b6a7-dfaf-4c10-88be-f482a5c58667" containerName="extract-utilities" Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.016804 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bc1b6a7-dfaf-4c10-88be-f482a5c58667" containerName="registry-server" Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.018779 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vqtpd" Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.022432 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vqtpd"] Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.181062 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sctnn\" (UniqueName: \"kubernetes.io/projected/8aaa9295-db53-415c-a01d-80328833af1c-kube-api-access-sctnn\") pod \"community-operators-vqtpd\" (UID: \"8aaa9295-db53-415c-a01d-80328833af1c\") " pod="openshift-marketplace/community-operators-vqtpd" Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.181154 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8aaa9295-db53-415c-a01d-80328833af1c-catalog-content\") pod \"community-operators-vqtpd\" (UID: \"8aaa9295-db53-415c-a01d-80328833af1c\") " pod="openshift-marketplace/community-operators-vqtpd" Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.181262 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8aaa9295-db53-415c-a01d-80328833af1c-utilities\") pod \"community-operators-vqtpd\" (UID: \"8aaa9295-db53-415c-a01d-80328833af1c\") " pod="openshift-marketplace/community-operators-vqtpd" Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.283298 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8aaa9295-db53-415c-a01d-80328833af1c-utilities\") pod \"community-operators-vqtpd\" (UID: \"8aaa9295-db53-415c-a01d-80328833af1c\") " pod="openshift-marketplace/community-operators-vqtpd" Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.283404 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sctnn\" (UniqueName: \"kubernetes.io/projected/8aaa9295-db53-415c-a01d-80328833af1c-kube-api-access-sctnn\") pod \"community-operators-vqtpd\" (UID: \"8aaa9295-db53-415c-a01d-80328833af1c\") " pod="openshift-marketplace/community-operators-vqtpd" Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.283469 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8aaa9295-db53-415c-a01d-80328833af1c-catalog-content\") pod \"community-operators-vqtpd\" (UID: \"8aaa9295-db53-415c-a01d-80328833af1c\") " pod="openshift-marketplace/community-operators-vqtpd" Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.283994 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8aaa9295-db53-415c-a01d-80328833af1c-catalog-content\") pod \"community-operators-vqtpd\" (UID: \"8aaa9295-db53-415c-a01d-80328833af1c\") " pod="openshift-marketplace/community-operators-vqtpd" Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.284000 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8aaa9295-db53-415c-a01d-80328833af1c-utilities\") pod \"community-operators-vqtpd\" (UID: \"8aaa9295-db53-415c-a01d-80328833af1c\") " pod="openshift-marketplace/community-operators-vqtpd" Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.305828 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sctnn\" (UniqueName: \"kubernetes.io/projected/8aaa9295-db53-415c-a01d-80328833af1c-kube-api-access-sctnn\") pod \"community-operators-vqtpd\" (UID: \"8aaa9295-db53-415c-a01d-80328833af1c\") " pod="openshift-marketplace/community-operators-vqtpd" Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.401062 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vqtpd" Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.928068 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vqtpd"] Feb 20 08:44:00 crc kubenswrapper[4948]: I0220 08:44:00.497833 4948 generic.go:334] "Generic (PLEG): container finished" podID="8aaa9295-db53-415c-a01d-80328833af1c" containerID="04150064869f7e2104b08b6a1ea4042774631d5a1e64200e7d0b648e85fc12c8" exitCode=0 Feb 20 08:44:00 crc kubenswrapper[4948]: I0220 08:44:00.497934 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vqtpd" event={"ID":"8aaa9295-db53-415c-a01d-80328833af1c","Type":"ContainerDied","Data":"04150064869f7e2104b08b6a1ea4042774631d5a1e64200e7d0b648e85fc12c8"} Feb 20 08:44:00 crc kubenswrapper[4948]: I0220 08:44:00.498410 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vqtpd" event={"ID":"8aaa9295-db53-415c-a01d-80328833af1c","Type":"ContainerStarted","Data":"35e21374dff7e8f3d55b2da70cd138e5ac76100c8db55005f609ee420971b833"} Feb 20 08:44:00 crc kubenswrapper[4948]: I0220 08:44:00.501557 4948 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 08:44:02 crc kubenswrapper[4948]: I0220 08:44:02.524201 4948 generic.go:334] "Generic (PLEG): container finished" podID="8aaa9295-db53-415c-a01d-80328833af1c" containerID="6043e2e2a61ed40b0ced1f92a7c96922825a7177951ad0332aa05a8120ec9df2" exitCode=0 Feb 20 08:44:02 crc kubenswrapper[4948]: I0220 08:44:02.524271 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vqtpd" event={"ID":"8aaa9295-db53-415c-a01d-80328833af1c","Type":"ContainerDied","Data":"6043e2e2a61ed40b0ced1f92a7c96922825a7177951ad0332aa05a8120ec9df2"} Feb 20 08:44:03 crc kubenswrapper[4948]: I0220 08:44:03.538082 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vqtpd" event={"ID":"8aaa9295-db53-415c-a01d-80328833af1c","Type":"ContainerStarted","Data":"1c1738fd62956b1cf907c7bb63dad6c36c4b8e5f2086bdcaa6aedd2a9f973732"} Feb 20 08:44:03 crc kubenswrapper[4948]: I0220 08:44:03.577632 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vqtpd" podStartSLOduration=3.069184415 podStartE2EDuration="5.577610884s" podCreationTimestamp="2026-02-20 08:43:58 +0000 UTC" firstStartedPulling="2026-02-20 08:44:00.500778716 +0000 UTC m=+2289.475273586" lastFinishedPulling="2026-02-20 08:44:03.009205195 +0000 UTC m=+2291.983700055" observedRunningTime="2026-02-20 08:44:03.568305543 +0000 UTC m=+2292.542800403" watchObservedRunningTime="2026-02-20 08:44:03.577610884 +0000 UTC m=+2292.552105714" Feb 20 08:44:07 crc kubenswrapper[4948]: I0220 08:44:07.719123 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-52bhl"] Feb 20 08:44:07 crc kubenswrapper[4948]: I0220 08:44:07.721554 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-52bhl" Feb 20 08:44:07 crc kubenswrapper[4948]: I0220 08:44:07.754247 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-52bhl"] Feb 20 08:44:07 crc kubenswrapper[4948]: I0220 08:44:07.894718 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chz94\" (UniqueName: \"kubernetes.io/projected/7e320036-f16c-4305-8591-c2f37e560ae1-kube-api-access-chz94\") pod \"certified-operators-52bhl\" (UID: \"7e320036-f16c-4305-8591-c2f37e560ae1\") " pod="openshift-marketplace/certified-operators-52bhl" Feb 20 08:44:07 crc kubenswrapper[4948]: I0220 08:44:07.894784 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e320036-f16c-4305-8591-c2f37e560ae1-utilities\") pod \"certified-operators-52bhl\" (UID: \"7e320036-f16c-4305-8591-c2f37e560ae1\") " pod="openshift-marketplace/certified-operators-52bhl" Feb 20 08:44:07 crc kubenswrapper[4948]: I0220 08:44:07.894833 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e320036-f16c-4305-8591-c2f37e560ae1-catalog-content\") pod \"certified-operators-52bhl\" (UID: \"7e320036-f16c-4305-8591-c2f37e560ae1\") " pod="openshift-marketplace/certified-operators-52bhl" Feb 20 08:44:07 crc kubenswrapper[4948]: I0220 08:44:07.997194 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e320036-f16c-4305-8591-c2f37e560ae1-utilities\") pod \"certified-operators-52bhl\" (UID: \"7e320036-f16c-4305-8591-c2f37e560ae1\") " pod="openshift-marketplace/certified-operators-52bhl" Feb 20 08:44:07 crc kubenswrapper[4948]: I0220 08:44:07.997284 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e320036-f16c-4305-8591-c2f37e560ae1-catalog-content\") pod \"certified-operators-52bhl\" (UID: \"7e320036-f16c-4305-8591-c2f37e560ae1\") " pod="openshift-marketplace/certified-operators-52bhl" Feb 20 08:44:07 crc kubenswrapper[4948]: I0220 08:44:07.997632 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chz94\" (UniqueName: \"kubernetes.io/projected/7e320036-f16c-4305-8591-c2f37e560ae1-kube-api-access-chz94\") pod \"certified-operators-52bhl\" (UID: \"7e320036-f16c-4305-8591-c2f37e560ae1\") " pod="openshift-marketplace/certified-operators-52bhl" Feb 20 08:44:07 crc kubenswrapper[4948]: I0220 08:44:07.997780 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e320036-f16c-4305-8591-c2f37e560ae1-catalog-content\") pod \"certified-operators-52bhl\" (UID: \"7e320036-f16c-4305-8591-c2f37e560ae1\") " pod="openshift-marketplace/certified-operators-52bhl" Feb 20 08:44:07 crc kubenswrapper[4948]: I0220 08:44:07.997875 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e320036-f16c-4305-8591-c2f37e560ae1-utilities\") pod \"certified-operators-52bhl\" (UID: \"7e320036-f16c-4305-8591-c2f37e560ae1\") " pod="openshift-marketplace/certified-operators-52bhl" Feb 20 08:44:08 crc kubenswrapper[4948]: I0220 08:44:08.027453 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chz94\" (UniqueName: \"kubernetes.io/projected/7e320036-f16c-4305-8591-c2f37e560ae1-kube-api-access-chz94\") pod \"certified-operators-52bhl\" (UID: \"7e320036-f16c-4305-8591-c2f37e560ae1\") " pod="openshift-marketplace/certified-operators-52bhl" Feb 20 08:44:08 crc kubenswrapper[4948]: I0220 08:44:08.057277 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-52bhl" Feb 20 08:44:08 crc kubenswrapper[4948]: I0220 08:44:08.588889 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-52bhl"] Feb 20 08:44:08 crc kubenswrapper[4948]: W0220 08:44:08.602702 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e320036_f16c_4305_8591_c2f37e560ae1.slice/crio-36c7b6d8b8c2c363ef13c4f00576bc6b814e0270caaf06486e95056f5e0456ca WatchSource:0}: Error finding container 36c7b6d8b8c2c363ef13c4f00576bc6b814e0270caaf06486e95056f5e0456ca: Status 404 returned error can't find the container with id 36c7b6d8b8c2c363ef13c4f00576bc6b814e0270caaf06486e95056f5e0456ca Feb 20 08:44:09 crc kubenswrapper[4948]: I0220 08:44:09.401421 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vqtpd" Feb 20 08:44:09 crc kubenswrapper[4948]: I0220 08:44:09.401825 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vqtpd" Feb 20 08:44:09 crc kubenswrapper[4948]: I0220 08:44:09.468183 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vqtpd" Feb 20 08:44:09 crc kubenswrapper[4948]: I0220 08:44:09.602952 4948 generic.go:334] "Generic (PLEG): container finished" podID="7e320036-f16c-4305-8591-c2f37e560ae1" containerID="b6344b3659b55c85ab17bb84f2d75f533dcdaf801d948da41f4dce7365aba5b2" exitCode=0 Feb 20 08:44:09 crc kubenswrapper[4948]: I0220 08:44:09.606222 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-52bhl" event={"ID":"7e320036-f16c-4305-8591-c2f37e560ae1","Type":"ContainerDied","Data":"b6344b3659b55c85ab17bb84f2d75f533dcdaf801d948da41f4dce7365aba5b2"} Feb 20 08:44:09 crc kubenswrapper[4948]: I0220 08:44:09.606282 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-52bhl" event={"ID":"7e320036-f16c-4305-8591-c2f37e560ae1","Type":"ContainerStarted","Data":"36c7b6d8b8c2c363ef13c4f00576bc6b814e0270caaf06486e95056f5e0456ca"} Feb 20 08:44:09 crc kubenswrapper[4948]: I0220 08:44:09.671254 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vqtpd" Feb 20 08:44:09 crc kubenswrapper[4948]: I0220 08:44:09.723145 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:44:09 crc kubenswrapper[4948]: E0220 08:44:09.723483 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:44:10 crc kubenswrapper[4948]: I0220 08:44:10.612352 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-52bhl" event={"ID":"7e320036-f16c-4305-8591-c2f37e560ae1","Type":"ContainerStarted","Data":"183389c9a096dfb452e747af5b860f85837fe8aad0eeae056715142ef6fa86d6"} Feb 20 08:44:11 crc kubenswrapper[4948]: I0220 08:44:11.645382 4948 generic.go:334] "Generic (PLEG): container finished" podID="7e320036-f16c-4305-8591-c2f37e560ae1" containerID="183389c9a096dfb452e747af5b860f85837fe8aad0eeae056715142ef6fa86d6" exitCode=0 Feb 20 08:44:11 crc kubenswrapper[4948]: I0220 08:44:11.645440 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-52bhl" event={"ID":"7e320036-f16c-4305-8591-c2f37e560ae1","Type":"ContainerDied","Data":"183389c9a096dfb452e747af5b860f85837fe8aad0eeae056715142ef6fa86d6"} Feb 20 08:44:11 crc kubenswrapper[4948]: I0220 08:44:11.893545 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vqtpd"] Feb 20 08:44:11 crc kubenswrapper[4948]: I0220 08:44:11.894105 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vqtpd" podUID="8aaa9295-db53-415c-a01d-80328833af1c" containerName="registry-server" containerID="cri-o://1c1738fd62956b1cf907c7bb63dad6c36c4b8e5f2086bdcaa6aedd2a9f973732" gracePeriod=2 Feb 20 08:44:12 crc kubenswrapper[4948]: I0220 08:44:12.655454 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-52bhl" event={"ID":"7e320036-f16c-4305-8591-c2f37e560ae1","Type":"ContainerStarted","Data":"8dfd7a07c4d2e10fac00ea5e02e285c69b8c28107d28b2e8147168697f405036"} Feb 20 08:44:12 crc kubenswrapper[4948]: I0220 08:44:12.679539 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-52bhl" podStartSLOduration=3.245298351 podStartE2EDuration="5.679512661s" podCreationTimestamp="2026-02-20 08:44:07 +0000 UTC" firstStartedPulling="2026-02-20 08:44:09.609229686 +0000 UTC m=+2298.583724526" lastFinishedPulling="2026-02-20 08:44:12.043444016 +0000 UTC m=+2301.017938836" observedRunningTime="2026-02-20 08:44:12.673154614 +0000 UTC m=+2301.647649444" watchObservedRunningTime="2026-02-20 08:44:12.679512661 +0000 UTC m=+2301.654007481" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.569205 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vqtpd" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.666147 4948 generic.go:334] "Generic (PLEG): container finished" podID="8aaa9295-db53-415c-a01d-80328833af1c" containerID="1c1738fd62956b1cf907c7bb63dad6c36c4b8e5f2086bdcaa6aedd2a9f973732" exitCode=0 Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.666573 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vqtpd" event={"ID":"8aaa9295-db53-415c-a01d-80328833af1c","Type":"ContainerDied","Data":"1c1738fd62956b1cf907c7bb63dad6c36c4b8e5f2086bdcaa6aedd2a9f973732"} Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.666676 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vqtpd" event={"ID":"8aaa9295-db53-415c-a01d-80328833af1c","Type":"ContainerDied","Data":"35e21374dff7e8f3d55b2da70cd138e5ac76100c8db55005f609ee420971b833"} Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.666695 4948 scope.go:117] "RemoveContainer" containerID="1c1738fd62956b1cf907c7bb63dad6c36c4b8e5f2086bdcaa6aedd2a9f973732" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.666779 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vqtpd" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.705656 4948 scope.go:117] "RemoveContainer" containerID="6043e2e2a61ed40b0ced1f92a7c96922825a7177951ad0332aa05a8120ec9df2" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.715306 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sctnn\" (UniqueName: \"kubernetes.io/projected/8aaa9295-db53-415c-a01d-80328833af1c-kube-api-access-sctnn\") pod \"8aaa9295-db53-415c-a01d-80328833af1c\" (UID: \"8aaa9295-db53-415c-a01d-80328833af1c\") " Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.715498 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8aaa9295-db53-415c-a01d-80328833af1c-utilities\") pod \"8aaa9295-db53-415c-a01d-80328833af1c\" (UID: \"8aaa9295-db53-415c-a01d-80328833af1c\") " Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.715636 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8aaa9295-db53-415c-a01d-80328833af1c-catalog-content\") pod \"8aaa9295-db53-415c-a01d-80328833af1c\" (UID: \"8aaa9295-db53-415c-a01d-80328833af1c\") " Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.721834 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8aaa9295-db53-415c-a01d-80328833af1c-utilities" (OuterVolumeSpecName: "utilities") pod "8aaa9295-db53-415c-a01d-80328833af1c" (UID: "8aaa9295-db53-415c-a01d-80328833af1c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.725227 4948 scope.go:117] "RemoveContainer" containerID="04150064869f7e2104b08b6a1ea4042774631d5a1e64200e7d0b648e85fc12c8" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.727216 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8aaa9295-db53-415c-a01d-80328833af1c-kube-api-access-sctnn" (OuterVolumeSpecName: "kube-api-access-sctnn") pod "8aaa9295-db53-415c-a01d-80328833af1c" (UID: "8aaa9295-db53-415c-a01d-80328833af1c"). InnerVolumeSpecName "kube-api-access-sctnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.775077 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8aaa9295-db53-415c-a01d-80328833af1c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8aaa9295-db53-415c-a01d-80328833af1c" (UID: "8aaa9295-db53-415c-a01d-80328833af1c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.818405 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sctnn\" (UniqueName: \"kubernetes.io/projected/8aaa9295-db53-415c-a01d-80328833af1c-kube-api-access-sctnn\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.818440 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8aaa9295-db53-415c-a01d-80328833af1c-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.818449 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8aaa9295-db53-415c-a01d-80328833af1c-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.821090 4948 scope.go:117] "RemoveContainer" containerID="1c1738fd62956b1cf907c7bb63dad6c36c4b8e5f2086bdcaa6aedd2a9f973732" Feb 20 08:44:13 crc kubenswrapper[4948]: E0220 08:44:13.821760 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c1738fd62956b1cf907c7bb63dad6c36c4b8e5f2086bdcaa6aedd2a9f973732\": container with ID starting with 1c1738fd62956b1cf907c7bb63dad6c36c4b8e5f2086bdcaa6aedd2a9f973732 not found: ID does not exist" containerID="1c1738fd62956b1cf907c7bb63dad6c36c4b8e5f2086bdcaa6aedd2a9f973732" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.821801 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c1738fd62956b1cf907c7bb63dad6c36c4b8e5f2086bdcaa6aedd2a9f973732"} err="failed to get container status \"1c1738fd62956b1cf907c7bb63dad6c36c4b8e5f2086bdcaa6aedd2a9f973732\": rpc error: code = NotFound desc = could not find container \"1c1738fd62956b1cf907c7bb63dad6c36c4b8e5f2086bdcaa6aedd2a9f973732\": container with ID starting with 1c1738fd62956b1cf907c7bb63dad6c36c4b8e5f2086bdcaa6aedd2a9f973732 not found: ID does not exist" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.821826 4948 scope.go:117] "RemoveContainer" containerID="6043e2e2a61ed40b0ced1f92a7c96922825a7177951ad0332aa05a8120ec9df2" Feb 20 08:44:13 crc kubenswrapper[4948]: E0220 08:44:13.822107 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6043e2e2a61ed40b0ced1f92a7c96922825a7177951ad0332aa05a8120ec9df2\": container with ID starting with 6043e2e2a61ed40b0ced1f92a7c96922825a7177951ad0332aa05a8120ec9df2 not found: ID does not exist" containerID="6043e2e2a61ed40b0ced1f92a7c96922825a7177951ad0332aa05a8120ec9df2" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.822145 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6043e2e2a61ed40b0ced1f92a7c96922825a7177951ad0332aa05a8120ec9df2"} err="failed to get container status \"6043e2e2a61ed40b0ced1f92a7c96922825a7177951ad0332aa05a8120ec9df2\": rpc error: code = NotFound desc = could not find container \"6043e2e2a61ed40b0ced1f92a7c96922825a7177951ad0332aa05a8120ec9df2\": container with ID starting with 6043e2e2a61ed40b0ced1f92a7c96922825a7177951ad0332aa05a8120ec9df2 not found: ID does not exist" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.822169 4948 scope.go:117] "RemoveContainer" containerID="04150064869f7e2104b08b6a1ea4042774631d5a1e64200e7d0b648e85fc12c8" Feb 20 08:44:13 crc kubenswrapper[4948]: E0220 08:44:13.822412 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04150064869f7e2104b08b6a1ea4042774631d5a1e64200e7d0b648e85fc12c8\": container with ID starting with 04150064869f7e2104b08b6a1ea4042774631d5a1e64200e7d0b648e85fc12c8 not found: ID does not exist" containerID="04150064869f7e2104b08b6a1ea4042774631d5a1e64200e7d0b648e85fc12c8" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.822434 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04150064869f7e2104b08b6a1ea4042774631d5a1e64200e7d0b648e85fc12c8"} err="failed to get container status \"04150064869f7e2104b08b6a1ea4042774631d5a1e64200e7d0b648e85fc12c8\": rpc error: code = NotFound desc = could not find container \"04150064869f7e2104b08b6a1ea4042774631d5a1e64200e7d0b648e85fc12c8\": container with ID starting with 04150064869f7e2104b08b6a1ea4042774631d5a1e64200e7d0b648e85fc12c8 not found: ID does not exist" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.996270 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vqtpd"] Feb 20 08:44:14 crc kubenswrapper[4948]: I0220 08:44:14.003911 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vqtpd"] Feb 20 08:44:15 crc kubenswrapper[4948]: I0220 08:44:15.741586 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8aaa9295-db53-415c-a01d-80328833af1c" path="/var/lib/kubelet/pods/8aaa9295-db53-415c-a01d-80328833af1c/volumes" Feb 20 08:44:18 crc kubenswrapper[4948]: I0220 08:44:18.058468 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-52bhl" Feb 20 08:44:18 crc kubenswrapper[4948]: I0220 08:44:18.058756 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-52bhl" Feb 20 08:44:18 crc kubenswrapper[4948]: I0220 08:44:18.133099 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-52bhl" Feb 20 08:44:18 crc kubenswrapper[4948]: I0220 08:44:18.779966 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-52bhl" Feb 20 08:44:19 crc kubenswrapper[4948]: I0220 08:44:19.296639 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-52bhl"] Feb 20 08:44:20 crc kubenswrapper[4948]: I0220 08:44:20.750641 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-52bhl" podUID="7e320036-f16c-4305-8591-c2f37e560ae1" containerName="registry-server" containerID="cri-o://8dfd7a07c4d2e10fac00ea5e02e285c69b8c28107d28b2e8147168697f405036" gracePeriod=2 Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.197570 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-52bhl" Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.382533 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e320036-f16c-4305-8591-c2f37e560ae1-utilities\") pod \"7e320036-f16c-4305-8591-c2f37e560ae1\" (UID: \"7e320036-f16c-4305-8591-c2f37e560ae1\") " Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.382588 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e320036-f16c-4305-8591-c2f37e560ae1-catalog-content\") pod \"7e320036-f16c-4305-8591-c2f37e560ae1\" (UID: \"7e320036-f16c-4305-8591-c2f37e560ae1\") " Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.382795 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chz94\" (UniqueName: \"kubernetes.io/projected/7e320036-f16c-4305-8591-c2f37e560ae1-kube-api-access-chz94\") pod \"7e320036-f16c-4305-8591-c2f37e560ae1\" (UID: \"7e320036-f16c-4305-8591-c2f37e560ae1\") " Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.383806 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e320036-f16c-4305-8591-c2f37e560ae1-utilities" (OuterVolumeSpecName: "utilities") pod "7e320036-f16c-4305-8591-c2f37e560ae1" (UID: "7e320036-f16c-4305-8591-c2f37e560ae1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.389246 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e320036-f16c-4305-8591-c2f37e560ae1-kube-api-access-chz94" (OuterVolumeSpecName: "kube-api-access-chz94") pod "7e320036-f16c-4305-8591-c2f37e560ae1" (UID: "7e320036-f16c-4305-8591-c2f37e560ae1"). InnerVolumeSpecName "kube-api-access-chz94". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.444021 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e320036-f16c-4305-8591-c2f37e560ae1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7e320036-f16c-4305-8591-c2f37e560ae1" (UID: "7e320036-f16c-4305-8591-c2f37e560ae1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.485278 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chz94\" (UniqueName: \"kubernetes.io/projected/7e320036-f16c-4305-8591-c2f37e560ae1-kube-api-access-chz94\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.485321 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e320036-f16c-4305-8591-c2f37e560ae1-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.485336 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e320036-f16c-4305-8591-c2f37e560ae1-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.732398 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:44:21 crc kubenswrapper[4948]: E0220 08:44:21.732915 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.764596 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-52bhl" Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.764640 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-52bhl" event={"ID":"7e320036-f16c-4305-8591-c2f37e560ae1","Type":"ContainerDied","Data":"8dfd7a07c4d2e10fac00ea5e02e285c69b8c28107d28b2e8147168697f405036"} Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.764692 4948 scope.go:117] "RemoveContainer" containerID="8dfd7a07c4d2e10fac00ea5e02e285c69b8c28107d28b2e8147168697f405036" Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.764539 4948 generic.go:334] "Generic (PLEG): container finished" podID="7e320036-f16c-4305-8591-c2f37e560ae1" containerID="8dfd7a07c4d2e10fac00ea5e02e285c69b8c28107d28b2e8147168697f405036" exitCode=0 Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.767544 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-52bhl" event={"ID":"7e320036-f16c-4305-8591-c2f37e560ae1","Type":"ContainerDied","Data":"36c7b6d8b8c2c363ef13c4f00576bc6b814e0270caaf06486e95056f5e0456ca"} Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.798501 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-52bhl"] Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.798544 4948 scope.go:117] "RemoveContainer" containerID="183389c9a096dfb452e747af5b860f85837fe8aad0eeae056715142ef6fa86d6" Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.807296 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-52bhl"] Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.829646 4948 scope.go:117] "RemoveContainer" containerID="b6344b3659b55c85ab17bb84f2d75f533dcdaf801d948da41f4dce7365aba5b2" Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.873201 4948 scope.go:117] "RemoveContainer" containerID="8dfd7a07c4d2e10fac00ea5e02e285c69b8c28107d28b2e8147168697f405036" Feb 20 08:44:21 crc kubenswrapper[4948]: E0220 08:44:21.873765 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dfd7a07c4d2e10fac00ea5e02e285c69b8c28107d28b2e8147168697f405036\": container with ID starting with 8dfd7a07c4d2e10fac00ea5e02e285c69b8c28107d28b2e8147168697f405036 not found: ID does not exist" containerID="8dfd7a07c4d2e10fac00ea5e02e285c69b8c28107d28b2e8147168697f405036" Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.873817 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dfd7a07c4d2e10fac00ea5e02e285c69b8c28107d28b2e8147168697f405036"} err="failed to get container status \"8dfd7a07c4d2e10fac00ea5e02e285c69b8c28107d28b2e8147168697f405036\": rpc error: code = NotFound desc = could not find container \"8dfd7a07c4d2e10fac00ea5e02e285c69b8c28107d28b2e8147168697f405036\": container with ID starting with 8dfd7a07c4d2e10fac00ea5e02e285c69b8c28107d28b2e8147168697f405036 not found: ID does not exist" Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.873851 4948 scope.go:117] "RemoveContainer" containerID="183389c9a096dfb452e747af5b860f85837fe8aad0eeae056715142ef6fa86d6" Feb 20 08:44:21 crc kubenswrapper[4948]: E0220 08:44:21.874244 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"183389c9a096dfb452e747af5b860f85837fe8aad0eeae056715142ef6fa86d6\": container with ID starting with 183389c9a096dfb452e747af5b860f85837fe8aad0eeae056715142ef6fa86d6 not found: ID does not exist" containerID="183389c9a096dfb452e747af5b860f85837fe8aad0eeae056715142ef6fa86d6" Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.874269 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"183389c9a096dfb452e747af5b860f85837fe8aad0eeae056715142ef6fa86d6"} err="failed to get container status \"183389c9a096dfb452e747af5b860f85837fe8aad0eeae056715142ef6fa86d6\": rpc error: code = NotFound desc = could not find container \"183389c9a096dfb452e747af5b860f85837fe8aad0eeae056715142ef6fa86d6\": container with ID starting with 183389c9a096dfb452e747af5b860f85837fe8aad0eeae056715142ef6fa86d6 not found: ID does not exist" Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.874283 4948 scope.go:117] "RemoveContainer" containerID="b6344b3659b55c85ab17bb84f2d75f533dcdaf801d948da41f4dce7365aba5b2" Feb 20 08:44:21 crc kubenswrapper[4948]: E0220 08:44:21.874902 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6344b3659b55c85ab17bb84f2d75f533dcdaf801d948da41f4dce7365aba5b2\": container with ID starting with b6344b3659b55c85ab17bb84f2d75f533dcdaf801d948da41f4dce7365aba5b2 not found: ID does not exist" containerID="b6344b3659b55c85ab17bb84f2d75f533dcdaf801d948da41f4dce7365aba5b2" Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.874942 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6344b3659b55c85ab17bb84f2d75f533dcdaf801d948da41f4dce7365aba5b2"} err="failed to get container status \"b6344b3659b55c85ab17bb84f2d75f533dcdaf801d948da41f4dce7365aba5b2\": rpc error: code = NotFound desc = could not find container \"b6344b3659b55c85ab17bb84f2d75f533dcdaf801d948da41f4dce7365aba5b2\": container with ID starting with b6344b3659b55c85ab17bb84f2d75f533dcdaf801d948da41f4dce7365aba5b2 not found: ID does not exist" Feb 20 08:44:23 crc kubenswrapper[4948]: I0220 08:44:23.733961 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e320036-f16c-4305-8591-c2f37e560ae1" path="/var/lib/kubelet/pods/7e320036-f16c-4305-8591-c2f37e560ae1/volumes" Feb 20 08:44:27 crc kubenswrapper[4948]: I0220 08:44:27.819833 4948 generic.go:334] "Generic (PLEG): container finished" podID="6005ded0-9e21-4908-bbf0-33c710ba4341" containerID="2f2f83e7b2bae4f94ae51a61eb3c4783c5606e273956b4e39c8926e563dd497a" exitCode=0 Feb 20 08:44:27 crc kubenswrapper[4948]: I0220 08:44:27.819914 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" event={"ID":"6005ded0-9e21-4908-bbf0-33c710ba4341","Type":"ContainerDied","Data":"2f2f83e7b2bae4f94ae51a61eb3c4783c5606e273956b4e39c8926e563dd497a"} Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.207805 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.352712 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-extra-config-0\") pod \"6005ded0-9e21-4908-bbf0-33c710ba4341\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.352751 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-1\") pod \"6005ded0-9e21-4908-bbf0-33c710ba4341\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.352778 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-migration-ssh-key-0\") pod \"6005ded0-9e21-4908-bbf0-33c710ba4341\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.352848 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ljgt\" (UniqueName: \"kubernetes.io/projected/6005ded0-9e21-4908-bbf0-33c710ba4341-kube-api-access-8ljgt\") pod \"6005ded0-9e21-4908-bbf0-33c710ba4341\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.352880 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-2\") pod \"6005ded0-9e21-4908-bbf0-33c710ba4341\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.352917 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-3\") pod \"6005ded0-9e21-4908-bbf0-33c710ba4341\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.352972 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-inventory\") pod \"6005ded0-9e21-4908-bbf0-33c710ba4341\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.353027 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-ssh-key-openstack-edpm-ipam\") pod \"6005ded0-9e21-4908-bbf0-33c710ba4341\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.353075 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-combined-ca-bundle\") pod \"6005ded0-9e21-4908-bbf0-33c710ba4341\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.353132 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-migration-ssh-key-1\") pod \"6005ded0-9e21-4908-bbf0-33c710ba4341\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.353173 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-0\") pod \"6005ded0-9e21-4908-bbf0-33c710ba4341\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.358377 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "6005ded0-9e21-4908-bbf0-33c710ba4341" (UID: "6005ded0-9e21-4908-bbf0-33c710ba4341"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.358805 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6005ded0-9e21-4908-bbf0-33c710ba4341-kube-api-access-8ljgt" (OuterVolumeSpecName: "kube-api-access-8ljgt") pod "6005ded0-9e21-4908-bbf0-33c710ba4341" (UID: "6005ded0-9e21-4908-bbf0-33c710ba4341"). InnerVolumeSpecName "kube-api-access-8ljgt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.386217 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-2" (OuterVolumeSpecName: "nova-cell1-compute-config-2") pod "6005ded0-9e21-4908-bbf0-33c710ba4341" (UID: "6005ded0-9e21-4908-bbf0-33c710ba4341"). InnerVolumeSpecName "nova-cell1-compute-config-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.388284 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "6005ded0-9e21-4908-bbf0-33c710ba4341" (UID: "6005ded0-9e21-4908-bbf0-33c710ba4341"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.390475 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "6005ded0-9e21-4908-bbf0-33c710ba4341" (UID: "6005ded0-9e21-4908-bbf0-33c710ba4341"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.391788 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "6005ded0-9e21-4908-bbf0-33c710ba4341" (UID: "6005ded0-9e21-4908-bbf0-33c710ba4341"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.398135 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "6005ded0-9e21-4908-bbf0-33c710ba4341" (UID: "6005ded0-9e21-4908-bbf0-33c710ba4341"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.398884 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-3" (OuterVolumeSpecName: "nova-cell1-compute-config-3") pod "6005ded0-9e21-4908-bbf0-33c710ba4341" (UID: "6005ded0-9e21-4908-bbf0-33c710ba4341"). InnerVolumeSpecName "nova-cell1-compute-config-3". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.400072 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "6005ded0-9e21-4908-bbf0-33c710ba4341" (UID: "6005ded0-9e21-4908-bbf0-33c710ba4341"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.400287 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "6005ded0-9e21-4908-bbf0-33c710ba4341" (UID: "6005ded0-9e21-4908-bbf0-33c710ba4341"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.422009 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-inventory" (OuterVolumeSpecName: "inventory") pod "6005ded0-9e21-4908-bbf0-33c710ba4341" (UID: "6005ded0-9e21-4908-bbf0-33c710ba4341"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.455711 4948 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.456090 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ljgt\" (UniqueName: \"kubernetes.io/projected/6005ded0-9e21-4908-bbf0-33c710ba4341-kube-api-access-8ljgt\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.456104 4948 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-2\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.456116 4948 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-3\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.456129 4948 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.456141 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.456152 4948 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.456165 4948 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.456176 4948 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.456191 4948 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.456203 4948 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.837003 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" event={"ID":"6005ded0-9e21-4908-bbf0-33c710ba4341","Type":"ContainerDied","Data":"f847f6d88e05df089793a5fb35fc239a9b46a7a0da87f3039554ac3950729267"} Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.837044 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f847f6d88e05df089793a5fb35fc239a9b46a7a0da87f3039554ac3950729267" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.837051 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.950217 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd"] Feb 20 08:44:29 crc kubenswrapper[4948]: E0220 08:44:29.950641 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8aaa9295-db53-415c-a01d-80328833af1c" containerName="registry-server" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.950665 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="8aaa9295-db53-415c-a01d-80328833af1c" containerName="registry-server" Feb 20 08:44:29 crc kubenswrapper[4948]: E0220 08:44:29.950679 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e320036-f16c-4305-8591-c2f37e560ae1" containerName="extract-utilities" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.950686 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e320036-f16c-4305-8591-c2f37e560ae1" containerName="extract-utilities" Feb 20 08:44:29 crc kubenswrapper[4948]: E0220 08:44:29.950698 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6005ded0-9e21-4908-bbf0-33c710ba4341" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.950706 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="6005ded0-9e21-4908-bbf0-33c710ba4341" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 20 08:44:29 crc kubenswrapper[4948]: E0220 08:44:29.950733 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e320036-f16c-4305-8591-c2f37e560ae1" containerName="extract-content" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.950740 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e320036-f16c-4305-8591-c2f37e560ae1" containerName="extract-content" Feb 20 08:44:29 crc kubenswrapper[4948]: E0220 08:44:29.950753 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8aaa9295-db53-415c-a01d-80328833af1c" containerName="extract-utilities" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.950760 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="8aaa9295-db53-415c-a01d-80328833af1c" containerName="extract-utilities" Feb 20 08:44:29 crc kubenswrapper[4948]: E0220 08:44:29.950773 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8aaa9295-db53-415c-a01d-80328833af1c" containerName="extract-content" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.950781 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="8aaa9295-db53-415c-a01d-80328833af1c" containerName="extract-content" Feb 20 08:44:29 crc kubenswrapper[4948]: E0220 08:44:29.950795 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e320036-f16c-4305-8591-c2f37e560ae1" containerName="registry-server" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.950802 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e320036-f16c-4305-8591-c2f37e560ae1" containerName="registry-server" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.951072 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e320036-f16c-4305-8591-c2f37e560ae1" containerName="registry-server" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.951090 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="6005ded0-9e21-4908-bbf0-33c710ba4341" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.951120 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="8aaa9295-db53-415c-a01d-80328833af1c" containerName="registry-server" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.951879 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.954019 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.954232 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.954356 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.954470 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sgx49" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.955171 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.959945 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd"] Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.065509 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xp6fd\" (UniqueName: \"kubernetes.io/projected/06ed5839-bc31-4691-8586-cd42c0413006-kube-api-access-xp6fd\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.065578 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.065603 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.065633 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.065653 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.065765 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.065846 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.167430 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.167481 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.167538 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.167560 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.167632 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xp6fd\" (UniqueName: \"kubernetes.io/projected/06ed5839-bc31-4691-8586-cd42c0413006-kube-api-access-xp6fd\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.167675 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.167694 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.175249 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.177595 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.178375 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.179031 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.183639 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.197623 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.205738 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xp6fd\" (UniqueName: \"kubernetes.io/projected/06ed5839-bc31-4691-8586-cd42c0413006-kube-api-access-xp6fd\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.267514 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.853187 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd"] Feb 20 08:44:30 crc kubenswrapper[4948]: W0220 08:44:30.858873 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod06ed5839_bc31_4691_8586_cd42c0413006.slice/crio-d456a9922236504054c207d73b58e4c58e11b56c6a928f1ce1badf2b7fbf1f58 WatchSource:0}: Error finding container d456a9922236504054c207d73b58e4c58e11b56c6a928f1ce1badf2b7fbf1f58: Status 404 returned error can't find the container with id d456a9922236504054c207d73b58e4c58e11b56c6a928f1ce1badf2b7fbf1f58 Feb 20 08:44:31 crc kubenswrapper[4948]: I0220 08:44:31.855663 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" event={"ID":"06ed5839-bc31-4691-8586-cd42c0413006","Type":"ContainerStarted","Data":"9f9449ceeed4dadc49534402fd6aefa2d193a1d53da8937c431d3883a09b7e37"} Feb 20 08:44:31 crc kubenswrapper[4948]: I0220 08:44:31.856034 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" event={"ID":"06ed5839-bc31-4691-8586-cd42c0413006","Type":"ContainerStarted","Data":"d456a9922236504054c207d73b58e4c58e11b56c6a928f1ce1badf2b7fbf1f58"} Feb 20 08:44:31 crc kubenswrapper[4948]: I0220 08:44:31.889971 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" podStartSLOduration=2.469003812 podStartE2EDuration="2.889945458s" podCreationTimestamp="2026-02-20 08:44:29 +0000 UTC" firstStartedPulling="2026-02-20 08:44:30.861749882 +0000 UTC m=+2319.836244722" lastFinishedPulling="2026-02-20 08:44:31.282691548 +0000 UTC m=+2320.257186368" observedRunningTime="2026-02-20 08:44:31.873701736 +0000 UTC m=+2320.848196556" watchObservedRunningTime="2026-02-20 08:44:31.889945458 +0000 UTC m=+2320.864440288" Feb 20 08:44:32 crc kubenswrapper[4948]: I0220 08:44:32.722470 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:44:32 crc kubenswrapper[4948]: E0220 08:44:32.722833 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:44:44 crc kubenswrapper[4948]: I0220 08:44:44.723270 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:44:44 crc kubenswrapper[4948]: E0220 08:44:44.724561 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:44:59 crc kubenswrapper[4948]: I0220 08:44:59.723576 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:44:59 crc kubenswrapper[4948]: E0220 08:44:59.724821 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:45:00 crc kubenswrapper[4948]: I0220 08:45:00.143717 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn"] Feb 20 08:45:00 crc kubenswrapper[4948]: I0220 08:45:00.145079 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn" Feb 20 08:45:00 crc kubenswrapper[4948]: I0220 08:45:00.148137 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 20 08:45:00 crc kubenswrapper[4948]: I0220 08:45:00.157927 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn"] Feb 20 08:45:00 crc kubenswrapper[4948]: I0220 08:45:00.158891 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 20 08:45:00 crc kubenswrapper[4948]: I0220 08:45:00.279300 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1b7070ce-cb29-445c-8f62-dd8b8db24431-config-volume\") pod \"collect-profiles-29526285-5lvgn\" (UID: \"1b7070ce-cb29-445c-8f62-dd8b8db24431\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn" Feb 20 08:45:00 crc kubenswrapper[4948]: I0220 08:45:00.279384 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sb977\" (UniqueName: \"kubernetes.io/projected/1b7070ce-cb29-445c-8f62-dd8b8db24431-kube-api-access-sb977\") pod \"collect-profiles-29526285-5lvgn\" (UID: \"1b7070ce-cb29-445c-8f62-dd8b8db24431\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn" Feb 20 08:45:00 crc kubenswrapper[4948]: I0220 08:45:00.279491 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1b7070ce-cb29-445c-8f62-dd8b8db24431-secret-volume\") pod \"collect-profiles-29526285-5lvgn\" (UID: \"1b7070ce-cb29-445c-8f62-dd8b8db24431\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn" Feb 20 08:45:00 crc kubenswrapper[4948]: I0220 08:45:00.381808 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sb977\" (UniqueName: \"kubernetes.io/projected/1b7070ce-cb29-445c-8f62-dd8b8db24431-kube-api-access-sb977\") pod \"collect-profiles-29526285-5lvgn\" (UID: \"1b7070ce-cb29-445c-8f62-dd8b8db24431\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn" Feb 20 08:45:00 crc kubenswrapper[4948]: I0220 08:45:00.382431 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1b7070ce-cb29-445c-8f62-dd8b8db24431-secret-volume\") pod \"collect-profiles-29526285-5lvgn\" (UID: \"1b7070ce-cb29-445c-8f62-dd8b8db24431\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn" Feb 20 08:45:00 crc kubenswrapper[4948]: I0220 08:45:00.382685 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1b7070ce-cb29-445c-8f62-dd8b8db24431-config-volume\") pod \"collect-profiles-29526285-5lvgn\" (UID: \"1b7070ce-cb29-445c-8f62-dd8b8db24431\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn" Feb 20 08:45:00 crc kubenswrapper[4948]: I0220 08:45:00.384257 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1b7070ce-cb29-445c-8f62-dd8b8db24431-config-volume\") pod \"collect-profiles-29526285-5lvgn\" (UID: \"1b7070ce-cb29-445c-8f62-dd8b8db24431\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn" Feb 20 08:45:00 crc kubenswrapper[4948]: I0220 08:45:00.393447 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1b7070ce-cb29-445c-8f62-dd8b8db24431-secret-volume\") pod \"collect-profiles-29526285-5lvgn\" (UID: \"1b7070ce-cb29-445c-8f62-dd8b8db24431\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn" Feb 20 08:45:00 crc kubenswrapper[4948]: I0220 08:45:00.405592 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sb977\" (UniqueName: \"kubernetes.io/projected/1b7070ce-cb29-445c-8f62-dd8b8db24431-kube-api-access-sb977\") pod \"collect-profiles-29526285-5lvgn\" (UID: \"1b7070ce-cb29-445c-8f62-dd8b8db24431\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn" Feb 20 08:45:00 crc kubenswrapper[4948]: I0220 08:45:00.463680 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn" Feb 20 08:45:00 crc kubenswrapper[4948]: I0220 08:45:00.958186 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn"] Feb 20 08:45:01 crc kubenswrapper[4948]: I0220 08:45:01.452611 4948 generic.go:334] "Generic (PLEG): container finished" podID="1b7070ce-cb29-445c-8f62-dd8b8db24431" containerID="0487ec6712045782955f0041f4583049c1ce40e10480a8b6d989918d4fa3552c" exitCode=0 Feb 20 08:45:01 crc kubenswrapper[4948]: I0220 08:45:01.452669 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn" event={"ID":"1b7070ce-cb29-445c-8f62-dd8b8db24431","Type":"ContainerDied","Data":"0487ec6712045782955f0041f4583049c1ce40e10480a8b6d989918d4fa3552c"} Feb 20 08:45:01 crc kubenswrapper[4948]: I0220 08:45:01.452992 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn" event={"ID":"1b7070ce-cb29-445c-8f62-dd8b8db24431","Type":"ContainerStarted","Data":"0f5ebdcd3104c719474346869a61ea8bbb58759dd8a0ed7585670e633c56ed16"} Feb 20 08:45:02 crc kubenswrapper[4948]: I0220 08:45:02.806025 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn" Feb 20 08:45:03 crc kubenswrapper[4948]: I0220 08:45:03.006985 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb977\" (UniqueName: \"kubernetes.io/projected/1b7070ce-cb29-445c-8f62-dd8b8db24431-kube-api-access-sb977\") pod \"1b7070ce-cb29-445c-8f62-dd8b8db24431\" (UID: \"1b7070ce-cb29-445c-8f62-dd8b8db24431\") " Feb 20 08:45:03 crc kubenswrapper[4948]: I0220 08:45:03.007471 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1b7070ce-cb29-445c-8f62-dd8b8db24431-config-volume\") pod \"1b7070ce-cb29-445c-8f62-dd8b8db24431\" (UID: \"1b7070ce-cb29-445c-8f62-dd8b8db24431\") " Feb 20 08:45:03 crc kubenswrapper[4948]: I0220 08:45:03.007694 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1b7070ce-cb29-445c-8f62-dd8b8db24431-secret-volume\") pod \"1b7070ce-cb29-445c-8f62-dd8b8db24431\" (UID: \"1b7070ce-cb29-445c-8f62-dd8b8db24431\") " Feb 20 08:45:03 crc kubenswrapper[4948]: I0220 08:45:03.007852 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b7070ce-cb29-445c-8f62-dd8b8db24431-config-volume" (OuterVolumeSpecName: "config-volume") pod "1b7070ce-cb29-445c-8f62-dd8b8db24431" (UID: "1b7070ce-cb29-445c-8f62-dd8b8db24431"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:45:03 crc kubenswrapper[4948]: I0220 08:45:03.008335 4948 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1b7070ce-cb29-445c-8f62-dd8b8db24431-config-volume\") on node \"crc\" DevicePath \"\"" Feb 20 08:45:03 crc kubenswrapper[4948]: I0220 08:45:03.012430 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b7070ce-cb29-445c-8f62-dd8b8db24431-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1b7070ce-cb29-445c-8f62-dd8b8db24431" (UID: "1b7070ce-cb29-445c-8f62-dd8b8db24431"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:45:03 crc kubenswrapper[4948]: I0220 08:45:03.012677 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b7070ce-cb29-445c-8f62-dd8b8db24431-kube-api-access-sb977" (OuterVolumeSpecName: "kube-api-access-sb977") pod "1b7070ce-cb29-445c-8f62-dd8b8db24431" (UID: "1b7070ce-cb29-445c-8f62-dd8b8db24431"). InnerVolumeSpecName "kube-api-access-sb977". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:45:03 crc kubenswrapper[4948]: I0220 08:45:03.109551 4948 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1b7070ce-cb29-445c-8f62-dd8b8db24431-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 20 08:45:03 crc kubenswrapper[4948]: I0220 08:45:03.109586 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb977\" (UniqueName: \"kubernetes.io/projected/1b7070ce-cb29-445c-8f62-dd8b8db24431-kube-api-access-sb977\") on node \"crc\" DevicePath \"\"" Feb 20 08:45:03 crc kubenswrapper[4948]: I0220 08:45:03.470889 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn" event={"ID":"1b7070ce-cb29-445c-8f62-dd8b8db24431","Type":"ContainerDied","Data":"0f5ebdcd3104c719474346869a61ea8bbb58759dd8a0ed7585670e633c56ed16"} Feb 20 08:45:03 crc kubenswrapper[4948]: I0220 08:45:03.471502 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f5ebdcd3104c719474346869a61ea8bbb58759dd8a0ed7585670e633c56ed16" Feb 20 08:45:03 crc kubenswrapper[4948]: I0220 08:45:03.470949 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn" Feb 20 08:45:03 crc kubenswrapper[4948]: E0220 08:45:03.675648 4948 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b7070ce_cb29_445c_8f62_dd8b8db24431.slice/crio-0f5ebdcd3104c719474346869a61ea8bbb58759dd8a0ed7585670e633c56ed16\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b7070ce_cb29_445c_8f62_dd8b8db24431.slice\": RecentStats: unable to find data in memory cache]" Feb 20 08:45:03 crc kubenswrapper[4948]: I0220 08:45:03.876027 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk"] Feb 20 08:45:03 crc kubenswrapper[4948]: I0220 08:45:03.884761 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk"] Feb 20 08:45:05 crc kubenswrapper[4948]: I0220 08:45:05.739187 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f77016a6-cd35-49df-b6b5-65b4858b41c9" path="/var/lib/kubelet/pods/f77016a6-cd35-49df-b6b5-65b4858b41c9/volumes" Feb 20 08:45:14 crc kubenswrapper[4948]: I0220 08:45:14.722273 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:45:14 crc kubenswrapper[4948]: E0220 08:45:14.723212 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:45:25 crc kubenswrapper[4948]: I0220 08:45:25.722965 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:45:25 crc kubenswrapper[4948]: E0220 08:45:25.723881 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:45:39 crc kubenswrapper[4948]: I0220 08:45:39.724226 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:45:39 crc kubenswrapper[4948]: E0220 08:45:39.725631 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:45:52 crc kubenswrapper[4948]: I0220 08:45:52.723539 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:45:52 crc kubenswrapper[4948]: E0220 08:45:52.724341 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:46:02 crc kubenswrapper[4948]: I0220 08:46:02.276139 4948 scope.go:117] "RemoveContainer" containerID="0520ff9e3a8f12a5228a352b59373978c12d5f10bb51923b12eab15acf564086" Feb 20 08:46:03 crc kubenswrapper[4948]: I0220 08:46:03.722625 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:46:03 crc kubenswrapper[4948]: E0220 08:46:03.723300 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:46:16 crc kubenswrapper[4948]: I0220 08:46:16.723164 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:46:16 crc kubenswrapper[4948]: E0220 08:46:16.724421 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:46:28 crc kubenswrapper[4948]: I0220 08:46:28.723158 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:46:28 crc kubenswrapper[4948]: E0220 08:46:28.724181 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:46:40 crc kubenswrapper[4948]: I0220 08:46:40.723243 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:46:40 crc kubenswrapper[4948]: E0220 08:46:40.725182 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:46:53 crc kubenswrapper[4948]: I0220 08:46:53.722930 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:46:53 crc kubenswrapper[4948]: E0220 08:46:53.723897 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:47:03 crc kubenswrapper[4948]: I0220 08:47:03.718470 4948 generic.go:334] "Generic (PLEG): container finished" podID="06ed5839-bc31-4691-8586-cd42c0413006" containerID="9f9449ceeed4dadc49534402fd6aefa2d193a1d53da8937c431d3883a09b7e37" exitCode=0 Feb 20 08:47:03 crc kubenswrapper[4948]: I0220 08:47:03.718569 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" event={"ID":"06ed5839-bc31-4691-8586-cd42c0413006","Type":"ContainerDied","Data":"9f9449ceeed4dadc49534402fd6aefa2d193a1d53da8937c431d3883a09b7e37"} Feb 20 08:47:04 crc kubenswrapper[4948]: I0220 08:47:04.723773 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:47:04 crc kubenswrapper[4948]: E0220 08:47:04.725157 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.230304 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.267211 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ssh-key-openstack-edpm-ipam\") pod \"06ed5839-bc31-4691-8586-cd42c0413006\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.267331 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-2\") pod \"06ed5839-bc31-4691-8586-cd42c0413006\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.267420 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-0\") pod \"06ed5839-bc31-4691-8586-cd42c0413006\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.267617 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-1\") pod \"06ed5839-bc31-4691-8586-cd42c0413006\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.267693 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-telemetry-combined-ca-bundle\") pod \"06ed5839-bc31-4691-8586-cd42c0413006\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.267749 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-inventory\") pod \"06ed5839-bc31-4691-8586-cd42c0413006\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.267896 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xp6fd\" (UniqueName: \"kubernetes.io/projected/06ed5839-bc31-4691-8586-cd42c0413006-kube-api-access-xp6fd\") pod \"06ed5839-bc31-4691-8586-cd42c0413006\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.274196 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06ed5839-bc31-4691-8586-cd42c0413006-kube-api-access-xp6fd" (OuterVolumeSpecName: "kube-api-access-xp6fd") pod "06ed5839-bc31-4691-8586-cd42c0413006" (UID: "06ed5839-bc31-4691-8586-cd42c0413006"). InnerVolumeSpecName "kube-api-access-xp6fd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.275575 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "06ed5839-bc31-4691-8586-cd42c0413006" (UID: "06ed5839-bc31-4691-8586-cd42c0413006"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.299851 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "06ed5839-bc31-4691-8586-cd42c0413006" (UID: "06ed5839-bc31-4691-8586-cd42c0413006"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.305411 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "06ed5839-bc31-4691-8586-cd42c0413006" (UID: "06ed5839-bc31-4691-8586-cd42c0413006"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.315899 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "06ed5839-bc31-4691-8586-cd42c0413006" (UID: "06ed5839-bc31-4691-8586-cd42c0413006"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.324335 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "06ed5839-bc31-4691-8586-cd42c0413006" (UID: "06ed5839-bc31-4691-8586-cd42c0413006"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.329622 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-inventory" (OuterVolumeSpecName: "inventory") pod "06ed5839-bc31-4691-8586-cd42c0413006" (UID: "06ed5839-bc31-4691-8586-cd42c0413006"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.370544 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.370573 4948 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.370585 4948 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.370595 4948 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.370603 4948 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.370615 4948 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.370623 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xp6fd\" (UniqueName: \"kubernetes.io/projected/06ed5839-bc31-4691-8586-cd42c0413006-kube-api-access-xp6fd\") on node \"crc\" DevicePath \"\"" Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.746104 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" event={"ID":"06ed5839-bc31-4691-8586-cd42c0413006","Type":"ContainerDied","Data":"d456a9922236504054c207d73b58e4c58e11b56c6a928f1ce1badf2b7fbf1f58"} Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.746158 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d456a9922236504054c207d73b58e4c58e11b56c6a928f1ce1badf2b7fbf1f58" Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.747217 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:47:16 crc kubenswrapper[4948]: I0220 08:47:16.739436 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:47:17 crc kubenswrapper[4948]: I0220 08:47:17.910560 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"dc74db389f6b091911817b21cddd65e37098545ff6852fc34fb4458a2808d454"} Feb 20 08:47:46 crc kubenswrapper[4948]: I0220 08:47:46.943047 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jbhz5"] Feb 20 08:47:46 crc kubenswrapper[4948]: E0220 08:47:46.945877 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06ed5839-bc31-4691-8586-cd42c0413006" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 20 08:47:46 crc kubenswrapper[4948]: I0220 08:47:46.946044 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="06ed5839-bc31-4691-8586-cd42c0413006" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 20 08:47:46 crc kubenswrapper[4948]: E0220 08:47:46.946153 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b7070ce-cb29-445c-8f62-dd8b8db24431" containerName="collect-profiles" Feb 20 08:47:46 crc kubenswrapper[4948]: I0220 08:47:46.946267 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b7070ce-cb29-445c-8f62-dd8b8db24431" containerName="collect-profiles" Feb 20 08:47:46 crc kubenswrapper[4948]: I0220 08:47:46.946752 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b7070ce-cb29-445c-8f62-dd8b8db24431" containerName="collect-profiles" Feb 20 08:47:46 crc kubenswrapper[4948]: I0220 08:47:46.955764 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="06ed5839-bc31-4691-8586-cd42c0413006" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 20 08:47:46 crc kubenswrapper[4948]: I0220 08:47:46.957836 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jbhz5" Feb 20 08:47:46 crc kubenswrapper[4948]: I0220 08:47:46.958331 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jbhz5"] Feb 20 08:47:46 crc kubenswrapper[4948]: I0220 08:47:46.991708 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10cca249-45b2-462c-a348-5881307f2f98-utilities\") pod \"redhat-marketplace-jbhz5\" (UID: \"10cca249-45b2-462c-a348-5881307f2f98\") " pod="openshift-marketplace/redhat-marketplace-jbhz5" Feb 20 08:47:46 crc kubenswrapper[4948]: I0220 08:47:46.994326 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10cca249-45b2-462c-a348-5881307f2f98-catalog-content\") pod \"redhat-marketplace-jbhz5\" (UID: \"10cca249-45b2-462c-a348-5881307f2f98\") " pod="openshift-marketplace/redhat-marketplace-jbhz5" Feb 20 08:47:46 crc kubenswrapper[4948]: I0220 08:47:46.994415 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7p69\" (UniqueName: \"kubernetes.io/projected/10cca249-45b2-462c-a348-5881307f2f98-kube-api-access-l7p69\") pod \"redhat-marketplace-jbhz5\" (UID: \"10cca249-45b2-462c-a348-5881307f2f98\") " pod="openshift-marketplace/redhat-marketplace-jbhz5" Feb 20 08:47:47 crc kubenswrapper[4948]: I0220 08:47:47.096370 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10cca249-45b2-462c-a348-5881307f2f98-utilities\") pod \"redhat-marketplace-jbhz5\" (UID: \"10cca249-45b2-462c-a348-5881307f2f98\") " pod="openshift-marketplace/redhat-marketplace-jbhz5" Feb 20 08:47:47 crc kubenswrapper[4948]: I0220 08:47:47.096482 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10cca249-45b2-462c-a348-5881307f2f98-catalog-content\") pod \"redhat-marketplace-jbhz5\" (UID: \"10cca249-45b2-462c-a348-5881307f2f98\") " pod="openshift-marketplace/redhat-marketplace-jbhz5" Feb 20 08:47:47 crc kubenswrapper[4948]: I0220 08:47:47.096510 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7p69\" (UniqueName: \"kubernetes.io/projected/10cca249-45b2-462c-a348-5881307f2f98-kube-api-access-l7p69\") pod \"redhat-marketplace-jbhz5\" (UID: \"10cca249-45b2-462c-a348-5881307f2f98\") " pod="openshift-marketplace/redhat-marketplace-jbhz5" Feb 20 08:47:47 crc kubenswrapper[4948]: I0220 08:47:47.097151 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10cca249-45b2-462c-a348-5881307f2f98-catalog-content\") pod \"redhat-marketplace-jbhz5\" (UID: \"10cca249-45b2-462c-a348-5881307f2f98\") " pod="openshift-marketplace/redhat-marketplace-jbhz5" Feb 20 08:47:47 crc kubenswrapper[4948]: I0220 08:47:47.097192 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10cca249-45b2-462c-a348-5881307f2f98-utilities\") pod \"redhat-marketplace-jbhz5\" (UID: \"10cca249-45b2-462c-a348-5881307f2f98\") " pod="openshift-marketplace/redhat-marketplace-jbhz5" Feb 20 08:47:47 crc kubenswrapper[4948]: I0220 08:47:47.118864 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7p69\" (UniqueName: \"kubernetes.io/projected/10cca249-45b2-462c-a348-5881307f2f98-kube-api-access-l7p69\") pod \"redhat-marketplace-jbhz5\" (UID: \"10cca249-45b2-462c-a348-5881307f2f98\") " pod="openshift-marketplace/redhat-marketplace-jbhz5" Feb 20 08:47:47 crc kubenswrapper[4948]: I0220 08:47:47.300411 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jbhz5" Feb 20 08:47:47 crc kubenswrapper[4948]: I0220 08:47:47.858333 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jbhz5"] Feb 20 08:47:48 crc kubenswrapper[4948]: I0220 08:47:48.236133 4948 generic.go:334] "Generic (PLEG): container finished" podID="10cca249-45b2-462c-a348-5881307f2f98" containerID="ed45a39cd5cc8d8919b2b3dce997a6b1812e40e55297096de6fbf28e3153a137" exitCode=0 Feb 20 08:47:48 crc kubenswrapper[4948]: I0220 08:47:48.236262 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jbhz5" event={"ID":"10cca249-45b2-462c-a348-5881307f2f98","Type":"ContainerDied","Data":"ed45a39cd5cc8d8919b2b3dce997a6b1812e40e55297096de6fbf28e3153a137"} Feb 20 08:47:48 crc kubenswrapper[4948]: I0220 08:47:48.236483 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jbhz5" event={"ID":"10cca249-45b2-462c-a348-5881307f2f98","Type":"ContainerStarted","Data":"2d7ce3d50e15df80a9a9553446d8695357118096db53c5e7f61047f7ee87b249"} Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.116464 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.119365 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.128304 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.128824 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.129420 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.130007 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-tbg4r" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.137656 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.140680 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.140720 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.140857 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.140915 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtx4h\" (UniqueName: \"kubernetes.io/projected/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-kube-api-access-dtx4h\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.141058 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.141088 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-config-data\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.141156 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.141188 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.141210 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.242821 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.242874 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtx4h\" (UniqueName: \"kubernetes.io/projected/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-kube-api-access-dtx4h\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.242955 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.242995 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-config-data\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.243046 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.243074 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.243099 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.243171 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.243192 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.243522 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.243567 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.245839 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-config-data\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.246671 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.250762 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.251085 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.252244 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jbhz5" event={"ID":"10cca249-45b2-462c-a348-5881307f2f98","Type":"ContainerStarted","Data":"0e3d35ab9b1e1ccdbcfb45233cef81d9c8c8ee3fc008c103fa0e76b0aabfaac1"} Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.253644 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.265622 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtx4h\" (UniqueName: \"kubernetes.io/projected/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-kube-api-access-dtx4h\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.271198 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.281633 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.453854 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.916990 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Feb 20 08:47:50 crc kubenswrapper[4948]: I0220 08:47:50.266415 4948 generic.go:334] "Generic (PLEG): container finished" podID="10cca249-45b2-462c-a348-5881307f2f98" containerID="0e3d35ab9b1e1ccdbcfb45233cef81d9c8c8ee3fc008c103fa0e76b0aabfaac1" exitCode=0 Feb 20 08:47:50 crc kubenswrapper[4948]: I0220 08:47:50.266475 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jbhz5" event={"ID":"10cca249-45b2-462c-a348-5881307f2f98","Type":"ContainerDied","Data":"0e3d35ab9b1e1ccdbcfb45233cef81d9c8c8ee3fc008c103fa0e76b0aabfaac1"} Feb 20 08:47:50 crc kubenswrapper[4948]: I0220 08:47:50.269887 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"0fefa0c4-ece9-4caf-8b41-8a64472c53e6","Type":"ContainerStarted","Data":"8d7b11cefa454a0dc17ba4434816118eb89ddebb724fafd4bee4a5317c66243c"} Feb 20 08:47:51 crc kubenswrapper[4948]: I0220 08:47:51.292188 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jbhz5" event={"ID":"10cca249-45b2-462c-a348-5881307f2f98","Type":"ContainerStarted","Data":"c4c8305352c501549986a399e8dfcfab3c6e25d6963a2311f98753f12586ab84"} Feb 20 08:47:51 crc kubenswrapper[4948]: I0220 08:47:51.315918 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jbhz5" podStartSLOduration=2.900921641 podStartE2EDuration="5.315902015s" podCreationTimestamp="2026-02-20 08:47:46 +0000 UTC" firstStartedPulling="2026-02-20 08:47:48.239431566 +0000 UTC m=+2517.213926436" lastFinishedPulling="2026-02-20 08:47:50.65441198 +0000 UTC m=+2519.628906810" observedRunningTime="2026-02-20 08:47:51.310796568 +0000 UTC m=+2520.285291388" watchObservedRunningTime="2026-02-20 08:47:51.315902015 +0000 UTC m=+2520.290396835" Feb 20 08:47:57 crc kubenswrapper[4948]: I0220 08:47:57.300609 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jbhz5" Feb 20 08:47:57 crc kubenswrapper[4948]: I0220 08:47:57.301218 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jbhz5" Feb 20 08:47:57 crc kubenswrapper[4948]: I0220 08:47:57.376148 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jbhz5" Feb 20 08:47:57 crc kubenswrapper[4948]: I0220 08:47:57.435625 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jbhz5" Feb 20 08:47:57 crc kubenswrapper[4948]: I0220 08:47:57.616536 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jbhz5"] Feb 20 08:47:59 crc kubenswrapper[4948]: I0220 08:47:59.357223 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jbhz5" podUID="10cca249-45b2-462c-a348-5881307f2f98" containerName="registry-server" containerID="cri-o://c4c8305352c501549986a399e8dfcfab3c6e25d6963a2311f98753f12586ab84" gracePeriod=2 Feb 20 08:47:59 crc kubenswrapper[4948]: I0220 08:47:59.886467 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jbhz5" Feb 20 08:48:00 crc kubenswrapper[4948]: I0220 08:48:00.065652 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7p69\" (UniqueName: \"kubernetes.io/projected/10cca249-45b2-462c-a348-5881307f2f98-kube-api-access-l7p69\") pod \"10cca249-45b2-462c-a348-5881307f2f98\" (UID: \"10cca249-45b2-462c-a348-5881307f2f98\") " Feb 20 08:48:00 crc kubenswrapper[4948]: I0220 08:48:00.065776 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10cca249-45b2-462c-a348-5881307f2f98-utilities\") pod \"10cca249-45b2-462c-a348-5881307f2f98\" (UID: \"10cca249-45b2-462c-a348-5881307f2f98\") " Feb 20 08:48:00 crc kubenswrapper[4948]: I0220 08:48:00.065922 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10cca249-45b2-462c-a348-5881307f2f98-catalog-content\") pod \"10cca249-45b2-462c-a348-5881307f2f98\" (UID: \"10cca249-45b2-462c-a348-5881307f2f98\") " Feb 20 08:48:00 crc kubenswrapper[4948]: I0220 08:48:00.066569 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10cca249-45b2-462c-a348-5881307f2f98-utilities" (OuterVolumeSpecName: "utilities") pod "10cca249-45b2-462c-a348-5881307f2f98" (UID: "10cca249-45b2-462c-a348-5881307f2f98"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:48:00 crc kubenswrapper[4948]: I0220 08:48:00.074783 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10cca249-45b2-462c-a348-5881307f2f98-kube-api-access-l7p69" (OuterVolumeSpecName: "kube-api-access-l7p69") pod "10cca249-45b2-462c-a348-5881307f2f98" (UID: "10cca249-45b2-462c-a348-5881307f2f98"). InnerVolumeSpecName "kube-api-access-l7p69". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:48:00 crc kubenswrapper[4948]: I0220 08:48:00.087848 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10cca249-45b2-462c-a348-5881307f2f98-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "10cca249-45b2-462c-a348-5881307f2f98" (UID: "10cca249-45b2-462c-a348-5881307f2f98"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:48:00 crc kubenswrapper[4948]: I0220 08:48:00.167838 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7p69\" (UniqueName: \"kubernetes.io/projected/10cca249-45b2-462c-a348-5881307f2f98-kube-api-access-l7p69\") on node \"crc\" DevicePath \"\"" Feb 20 08:48:00 crc kubenswrapper[4948]: I0220 08:48:00.167872 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10cca249-45b2-462c-a348-5881307f2f98-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:48:00 crc kubenswrapper[4948]: I0220 08:48:00.167907 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10cca249-45b2-462c-a348-5881307f2f98-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:48:00 crc kubenswrapper[4948]: I0220 08:48:00.369106 4948 generic.go:334] "Generic (PLEG): container finished" podID="10cca249-45b2-462c-a348-5881307f2f98" containerID="c4c8305352c501549986a399e8dfcfab3c6e25d6963a2311f98753f12586ab84" exitCode=0 Feb 20 08:48:00 crc kubenswrapper[4948]: I0220 08:48:00.369141 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jbhz5" Feb 20 08:48:00 crc kubenswrapper[4948]: I0220 08:48:00.369170 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jbhz5" event={"ID":"10cca249-45b2-462c-a348-5881307f2f98","Type":"ContainerDied","Data":"c4c8305352c501549986a399e8dfcfab3c6e25d6963a2311f98753f12586ab84"} Feb 20 08:48:00 crc kubenswrapper[4948]: I0220 08:48:00.369218 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jbhz5" event={"ID":"10cca249-45b2-462c-a348-5881307f2f98","Type":"ContainerDied","Data":"2d7ce3d50e15df80a9a9553446d8695357118096db53c5e7f61047f7ee87b249"} Feb 20 08:48:00 crc kubenswrapper[4948]: I0220 08:48:00.369247 4948 scope.go:117] "RemoveContainer" containerID="c4c8305352c501549986a399e8dfcfab3c6e25d6963a2311f98753f12586ab84" Feb 20 08:48:00 crc kubenswrapper[4948]: I0220 08:48:00.419312 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jbhz5"] Feb 20 08:48:00 crc kubenswrapper[4948]: I0220 08:48:00.426911 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jbhz5"] Feb 20 08:48:01 crc kubenswrapper[4948]: I0220 08:48:01.738277 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10cca249-45b2-462c-a348-5881307f2f98" path="/var/lib/kubelet/pods/10cca249-45b2-462c-a348-5881307f2f98/volumes" Feb 20 08:48:17 crc kubenswrapper[4948]: I0220 08:48:16.998822 4948 scope.go:117] "RemoveContainer" containerID="0e3d35ab9b1e1ccdbcfb45233cef81d9c8c8ee3fc008c103fa0e76b0aabfaac1" Feb 20 08:48:17 crc kubenswrapper[4948]: I0220 08:48:17.027640 4948 scope.go:117] "RemoveContainer" containerID="ed45a39cd5cc8d8919b2b3dce997a6b1812e40e55297096de6fbf28e3153a137" Feb 20 08:48:17 crc kubenswrapper[4948]: I0220 08:48:17.088309 4948 scope.go:117] "RemoveContainer" containerID="c4c8305352c501549986a399e8dfcfab3c6e25d6963a2311f98753f12586ab84" Feb 20 08:48:17 crc kubenswrapper[4948]: E0220 08:48:17.088791 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4c8305352c501549986a399e8dfcfab3c6e25d6963a2311f98753f12586ab84\": container with ID starting with c4c8305352c501549986a399e8dfcfab3c6e25d6963a2311f98753f12586ab84 not found: ID does not exist" containerID="c4c8305352c501549986a399e8dfcfab3c6e25d6963a2311f98753f12586ab84" Feb 20 08:48:17 crc kubenswrapper[4948]: I0220 08:48:17.088842 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4c8305352c501549986a399e8dfcfab3c6e25d6963a2311f98753f12586ab84"} err="failed to get container status \"c4c8305352c501549986a399e8dfcfab3c6e25d6963a2311f98753f12586ab84\": rpc error: code = NotFound desc = could not find container \"c4c8305352c501549986a399e8dfcfab3c6e25d6963a2311f98753f12586ab84\": container with ID starting with c4c8305352c501549986a399e8dfcfab3c6e25d6963a2311f98753f12586ab84 not found: ID does not exist" Feb 20 08:48:17 crc kubenswrapper[4948]: I0220 08:48:17.088873 4948 scope.go:117] "RemoveContainer" containerID="0e3d35ab9b1e1ccdbcfb45233cef81d9c8c8ee3fc008c103fa0e76b0aabfaac1" Feb 20 08:48:17 crc kubenswrapper[4948]: E0220 08:48:17.089402 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e3d35ab9b1e1ccdbcfb45233cef81d9c8c8ee3fc008c103fa0e76b0aabfaac1\": container with ID starting with 0e3d35ab9b1e1ccdbcfb45233cef81d9c8c8ee3fc008c103fa0e76b0aabfaac1 not found: ID does not exist" containerID="0e3d35ab9b1e1ccdbcfb45233cef81d9c8c8ee3fc008c103fa0e76b0aabfaac1" Feb 20 08:48:17 crc kubenswrapper[4948]: I0220 08:48:17.089439 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e3d35ab9b1e1ccdbcfb45233cef81d9c8c8ee3fc008c103fa0e76b0aabfaac1"} err="failed to get container status \"0e3d35ab9b1e1ccdbcfb45233cef81d9c8c8ee3fc008c103fa0e76b0aabfaac1\": rpc error: code = NotFound desc = could not find container \"0e3d35ab9b1e1ccdbcfb45233cef81d9c8c8ee3fc008c103fa0e76b0aabfaac1\": container with ID starting with 0e3d35ab9b1e1ccdbcfb45233cef81d9c8c8ee3fc008c103fa0e76b0aabfaac1 not found: ID does not exist" Feb 20 08:48:17 crc kubenswrapper[4948]: I0220 08:48:17.089464 4948 scope.go:117] "RemoveContainer" containerID="ed45a39cd5cc8d8919b2b3dce997a6b1812e40e55297096de6fbf28e3153a137" Feb 20 08:48:17 crc kubenswrapper[4948]: E0220 08:48:17.089782 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed45a39cd5cc8d8919b2b3dce997a6b1812e40e55297096de6fbf28e3153a137\": container with ID starting with ed45a39cd5cc8d8919b2b3dce997a6b1812e40e55297096de6fbf28e3153a137 not found: ID does not exist" containerID="ed45a39cd5cc8d8919b2b3dce997a6b1812e40e55297096de6fbf28e3153a137" Feb 20 08:48:17 crc kubenswrapper[4948]: I0220 08:48:17.089817 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed45a39cd5cc8d8919b2b3dce997a6b1812e40e55297096de6fbf28e3153a137"} err="failed to get container status \"ed45a39cd5cc8d8919b2b3dce997a6b1812e40e55297096de6fbf28e3153a137\": rpc error: code = NotFound desc = could not find container \"ed45a39cd5cc8d8919b2b3dce997a6b1812e40e55297096de6fbf28e3153a137\": container with ID starting with ed45a39cd5cc8d8919b2b3dce997a6b1812e40e55297096de6fbf28e3153a137 not found: ID does not exist" Feb 20 08:48:17 crc kubenswrapper[4948]: E0220 08:48:17.133850 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Feb 20 08:48:17 crc kubenswrapper[4948]: E0220 08:48:17.134067 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dtx4h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(0fefa0c4-ece9-4caf-8b41-8a64472c53e6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 08:48:17 crc kubenswrapper[4948]: E0220 08:48:17.136210 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="0fefa0c4-ece9-4caf-8b41-8a64472c53e6" Feb 20 08:48:17 crc kubenswrapper[4948]: E0220 08:48:17.574559 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="0fefa0c4-ece9-4caf-8b41-8a64472c53e6" Feb 20 08:48:30 crc kubenswrapper[4948]: I0220 08:48:30.151891 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Feb 20 08:48:31 crc kubenswrapper[4948]: I0220 08:48:31.756198 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"0fefa0c4-ece9-4caf-8b41-8a64472c53e6","Type":"ContainerStarted","Data":"7c1a082fb78288f20cddeb8b89d13f95d3e809472fa926ab7427ecd860311a72"} Feb 20 08:49:38 crc kubenswrapper[4948]: I0220 08:49:38.025615 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:49:38 crc kubenswrapper[4948]: I0220 08:49:38.026448 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:50:08 crc kubenswrapper[4948]: I0220 08:50:08.025603 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:50:08 crc kubenswrapper[4948]: I0220 08:50:08.027537 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:50:38 crc kubenswrapper[4948]: I0220 08:50:38.025103 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:50:38 crc kubenswrapper[4948]: I0220 08:50:38.025787 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:50:38 crc kubenswrapper[4948]: I0220 08:50:38.025850 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:50:38 crc kubenswrapper[4948]: I0220 08:50:38.026606 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dc74db389f6b091911817b21cddd65e37098545ff6852fc34fb4458a2808d454"} pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 08:50:38 crc kubenswrapper[4948]: I0220 08:50:38.026659 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" containerID="cri-o://dc74db389f6b091911817b21cddd65e37098545ff6852fc34fb4458a2808d454" gracePeriod=600 Feb 20 08:50:39 crc kubenswrapper[4948]: I0220 08:50:39.080594 4948 generic.go:334] "Generic (PLEG): container finished" podID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerID="dc74db389f6b091911817b21cddd65e37098545ff6852fc34fb4458a2808d454" exitCode=0 Feb 20 08:50:39 crc kubenswrapper[4948]: I0220 08:50:39.080657 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerDied","Data":"dc74db389f6b091911817b21cddd65e37098545ff6852fc34fb4458a2808d454"} Feb 20 08:50:39 crc kubenswrapper[4948]: I0220 08:50:39.081254 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5"} Feb 20 08:50:39 crc kubenswrapper[4948]: I0220 08:50:39.081286 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:50:39 crc kubenswrapper[4948]: I0220 08:50:39.106791 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=130.880422264 podStartE2EDuration="2m51.106776928s" podCreationTimestamp="2026-02-20 08:47:48 +0000 UTC" firstStartedPulling="2026-02-20 08:47:49.92155434 +0000 UTC m=+2518.896049160" lastFinishedPulling="2026-02-20 08:48:30.147908964 +0000 UTC m=+2559.122403824" observedRunningTime="2026-02-20 08:48:31.783936266 +0000 UTC m=+2560.758431126" watchObservedRunningTime="2026-02-20 08:50:39.106776928 +0000 UTC m=+2688.081271748" Feb 20 08:52:38 crc kubenswrapper[4948]: I0220 08:52:38.025018 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:52:38 crc kubenswrapper[4948]: I0220 08:52:38.025517 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:53:08 crc kubenswrapper[4948]: I0220 08:53:08.024760 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:53:08 crc kubenswrapper[4948]: I0220 08:53:08.025413 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:53:38 crc kubenswrapper[4948]: I0220 08:53:38.025245 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:53:38 crc kubenswrapper[4948]: I0220 08:53:38.026108 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:53:38 crc kubenswrapper[4948]: I0220 08:53:38.026191 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:53:38 crc kubenswrapper[4948]: I0220 08:53:38.027407 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5"} pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 08:53:38 crc kubenswrapper[4948]: I0220 08:53:38.027538 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" containerID="cri-o://cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" gracePeriod=600 Feb 20 08:53:38 crc kubenswrapper[4948]: E0220 08:53:38.149581 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:53:38 crc kubenswrapper[4948]: I0220 08:53:38.981616 4948 generic.go:334] "Generic (PLEG): container finished" podID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" exitCode=0 Feb 20 08:53:38 crc kubenswrapper[4948]: I0220 08:53:38.981689 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerDied","Data":"cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5"} Feb 20 08:53:38 crc kubenswrapper[4948]: I0220 08:53:38.982159 4948 scope.go:117] "RemoveContainer" containerID="dc74db389f6b091911817b21cddd65e37098545ff6852fc34fb4458a2808d454" Feb 20 08:53:38 crc kubenswrapper[4948]: I0220 08:53:38.983026 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:53:38 crc kubenswrapper[4948]: E0220 08:53:38.983480 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:53:44 crc kubenswrapper[4948]: I0220 08:53:44.820921 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dqcs5"] Feb 20 08:53:44 crc kubenswrapper[4948]: E0220 08:53:44.821788 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10cca249-45b2-462c-a348-5881307f2f98" containerName="extract-utilities" Feb 20 08:53:44 crc kubenswrapper[4948]: I0220 08:53:44.821802 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="10cca249-45b2-462c-a348-5881307f2f98" containerName="extract-utilities" Feb 20 08:53:44 crc kubenswrapper[4948]: E0220 08:53:44.821820 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10cca249-45b2-462c-a348-5881307f2f98" containerName="registry-server" Feb 20 08:53:44 crc kubenswrapper[4948]: I0220 08:53:44.821826 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="10cca249-45b2-462c-a348-5881307f2f98" containerName="registry-server" Feb 20 08:53:44 crc kubenswrapper[4948]: E0220 08:53:44.821845 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10cca249-45b2-462c-a348-5881307f2f98" containerName="extract-content" Feb 20 08:53:44 crc kubenswrapper[4948]: I0220 08:53:44.821852 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="10cca249-45b2-462c-a348-5881307f2f98" containerName="extract-content" Feb 20 08:53:44 crc kubenswrapper[4948]: I0220 08:53:44.822043 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="10cca249-45b2-462c-a348-5881307f2f98" containerName="registry-server" Feb 20 08:53:44 crc kubenswrapper[4948]: I0220 08:53:44.823364 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dqcs5" Feb 20 08:53:44 crc kubenswrapper[4948]: I0220 08:53:44.833353 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dqcs5"] Feb 20 08:53:44 crc kubenswrapper[4948]: I0220 08:53:44.925194 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-catalog-content\") pod \"redhat-operators-dqcs5\" (UID: \"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab\") " pod="openshift-marketplace/redhat-operators-dqcs5" Feb 20 08:53:44 crc kubenswrapper[4948]: I0220 08:53:44.925797 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q77f9\" (UniqueName: \"kubernetes.io/projected/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-kube-api-access-q77f9\") pod \"redhat-operators-dqcs5\" (UID: \"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab\") " pod="openshift-marketplace/redhat-operators-dqcs5" Feb 20 08:53:44 crc kubenswrapper[4948]: I0220 08:53:44.925995 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-utilities\") pod \"redhat-operators-dqcs5\" (UID: \"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab\") " pod="openshift-marketplace/redhat-operators-dqcs5" Feb 20 08:53:45 crc kubenswrapper[4948]: I0220 08:53:45.027621 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-catalog-content\") pod \"redhat-operators-dqcs5\" (UID: \"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab\") " pod="openshift-marketplace/redhat-operators-dqcs5" Feb 20 08:53:45 crc kubenswrapper[4948]: I0220 08:53:45.027740 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q77f9\" (UniqueName: \"kubernetes.io/projected/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-kube-api-access-q77f9\") pod \"redhat-operators-dqcs5\" (UID: \"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab\") " pod="openshift-marketplace/redhat-operators-dqcs5" Feb 20 08:53:45 crc kubenswrapper[4948]: I0220 08:53:45.027798 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-utilities\") pod \"redhat-operators-dqcs5\" (UID: \"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab\") " pod="openshift-marketplace/redhat-operators-dqcs5" Feb 20 08:53:45 crc kubenswrapper[4948]: I0220 08:53:45.028247 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-catalog-content\") pod \"redhat-operators-dqcs5\" (UID: \"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab\") " pod="openshift-marketplace/redhat-operators-dqcs5" Feb 20 08:53:45 crc kubenswrapper[4948]: I0220 08:53:45.028277 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-utilities\") pod \"redhat-operators-dqcs5\" (UID: \"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab\") " pod="openshift-marketplace/redhat-operators-dqcs5" Feb 20 08:53:45 crc kubenswrapper[4948]: I0220 08:53:45.053327 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q77f9\" (UniqueName: \"kubernetes.io/projected/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-kube-api-access-q77f9\") pod \"redhat-operators-dqcs5\" (UID: \"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab\") " pod="openshift-marketplace/redhat-operators-dqcs5" Feb 20 08:53:45 crc kubenswrapper[4948]: I0220 08:53:45.174177 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dqcs5" Feb 20 08:53:45 crc kubenswrapper[4948]: I0220 08:53:45.657316 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dqcs5"] Feb 20 08:53:46 crc kubenswrapper[4948]: I0220 08:53:46.045172 4948 generic.go:334] "Generic (PLEG): container finished" podID="5cb0bef2-1a21-46e7-905d-6dc6d0c50cab" containerID="e61abe7ff267f9bfeba63a25cfd6a81f5a070c5f25f48e0d6d8f167a049e5620" exitCode=0 Feb 20 08:53:46 crc kubenswrapper[4948]: I0220 08:53:46.045443 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqcs5" event={"ID":"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab","Type":"ContainerDied","Data":"e61abe7ff267f9bfeba63a25cfd6a81f5a070c5f25f48e0d6d8f167a049e5620"} Feb 20 08:53:46 crc kubenswrapper[4948]: I0220 08:53:46.045522 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqcs5" event={"ID":"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab","Type":"ContainerStarted","Data":"82329664b3d13e9d5f6f72bfbd0c9bfcb4cdedf3613453d9b8308754600e8fce"} Feb 20 08:53:46 crc kubenswrapper[4948]: I0220 08:53:46.047187 4948 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 08:53:48 crc kubenswrapper[4948]: I0220 08:53:48.066193 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqcs5" event={"ID":"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab","Type":"ContainerStarted","Data":"cdc48d4ad6f6ad6329e706c857c3263754fc1d230af2ba67afcff789d58a0f8f"} Feb 20 08:53:51 crc kubenswrapper[4948]: I0220 08:53:51.094799 4948 generic.go:334] "Generic (PLEG): container finished" podID="5cb0bef2-1a21-46e7-905d-6dc6d0c50cab" containerID="cdc48d4ad6f6ad6329e706c857c3263754fc1d230af2ba67afcff789d58a0f8f" exitCode=0 Feb 20 08:53:51 crc kubenswrapper[4948]: I0220 08:53:51.094899 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqcs5" event={"ID":"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab","Type":"ContainerDied","Data":"cdc48d4ad6f6ad6329e706c857c3263754fc1d230af2ba67afcff789d58a0f8f"} Feb 20 08:53:52 crc kubenswrapper[4948]: I0220 08:53:52.107331 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqcs5" event={"ID":"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab","Type":"ContainerStarted","Data":"b8438f9e17f81afa0b3fb46a37e2e751ee4f1773769441d7d00fc72148764d6d"} Feb 20 08:53:52 crc kubenswrapper[4948]: I0220 08:53:52.131110 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dqcs5" podStartSLOduration=2.700344155 podStartE2EDuration="8.131081336s" podCreationTimestamp="2026-02-20 08:53:44 +0000 UTC" firstStartedPulling="2026-02-20 08:53:46.046885546 +0000 UTC m=+2875.021380366" lastFinishedPulling="2026-02-20 08:53:51.477622727 +0000 UTC m=+2880.452117547" observedRunningTime="2026-02-20 08:53:52.128712518 +0000 UTC m=+2881.103207338" watchObservedRunningTime="2026-02-20 08:53:52.131081336 +0000 UTC m=+2881.105576186" Feb 20 08:53:52 crc kubenswrapper[4948]: I0220 08:53:52.722797 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:53:52 crc kubenswrapper[4948]: E0220 08:53:52.723173 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:53:55 crc kubenswrapper[4948]: I0220 08:53:55.175052 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dqcs5" Feb 20 08:53:55 crc kubenswrapper[4948]: I0220 08:53:55.175405 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dqcs5" Feb 20 08:53:56 crc kubenswrapper[4948]: I0220 08:53:56.280605 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dqcs5" podUID="5cb0bef2-1a21-46e7-905d-6dc6d0c50cab" containerName="registry-server" probeResult="failure" output=< Feb 20 08:53:56 crc kubenswrapper[4948]: timeout: failed to connect service ":50051" within 1s Feb 20 08:53:56 crc kubenswrapper[4948]: > Feb 20 08:53:58 crc kubenswrapper[4948]: I0220 08:53:58.230920 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7gx58"] Feb 20 08:53:58 crc kubenswrapper[4948]: I0220 08:53:58.234201 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7gx58" Feb 20 08:53:58 crc kubenswrapper[4948]: I0220 08:53:58.246393 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7gx58"] Feb 20 08:53:58 crc kubenswrapper[4948]: I0220 08:53:58.295489 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/862b1b6e-55b5-4038-97d3-c1d7bacd7952-catalog-content\") pod \"community-operators-7gx58\" (UID: \"862b1b6e-55b5-4038-97d3-c1d7bacd7952\") " pod="openshift-marketplace/community-operators-7gx58" Feb 20 08:53:58 crc kubenswrapper[4948]: I0220 08:53:58.295573 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtpqf\" (UniqueName: \"kubernetes.io/projected/862b1b6e-55b5-4038-97d3-c1d7bacd7952-kube-api-access-wtpqf\") pod \"community-operators-7gx58\" (UID: \"862b1b6e-55b5-4038-97d3-c1d7bacd7952\") " pod="openshift-marketplace/community-operators-7gx58" Feb 20 08:53:58 crc kubenswrapper[4948]: I0220 08:53:58.295764 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/862b1b6e-55b5-4038-97d3-c1d7bacd7952-utilities\") pod \"community-operators-7gx58\" (UID: \"862b1b6e-55b5-4038-97d3-c1d7bacd7952\") " pod="openshift-marketplace/community-operators-7gx58" Feb 20 08:53:58 crc kubenswrapper[4948]: I0220 08:53:58.397811 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/862b1b6e-55b5-4038-97d3-c1d7bacd7952-utilities\") pod \"community-operators-7gx58\" (UID: \"862b1b6e-55b5-4038-97d3-c1d7bacd7952\") " pod="openshift-marketplace/community-operators-7gx58" Feb 20 08:53:58 crc kubenswrapper[4948]: I0220 08:53:58.398112 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/862b1b6e-55b5-4038-97d3-c1d7bacd7952-catalog-content\") pod \"community-operators-7gx58\" (UID: \"862b1b6e-55b5-4038-97d3-c1d7bacd7952\") " pod="openshift-marketplace/community-operators-7gx58" Feb 20 08:53:58 crc kubenswrapper[4948]: I0220 08:53:58.398155 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtpqf\" (UniqueName: \"kubernetes.io/projected/862b1b6e-55b5-4038-97d3-c1d7bacd7952-kube-api-access-wtpqf\") pod \"community-operators-7gx58\" (UID: \"862b1b6e-55b5-4038-97d3-c1d7bacd7952\") " pod="openshift-marketplace/community-operators-7gx58" Feb 20 08:53:58 crc kubenswrapper[4948]: I0220 08:53:58.398710 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/862b1b6e-55b5-4038-97d3-c1d7bacd7952-utilities\") pod \"community-operators-7gx58\" (UID: \"862b1b6e-55b5-4038-97d3-c1d7bacd7952\") " pod="openshift-marketplace/community-operators-7gx58" Feb 20 08:53:58 crc kubenswrapper[4948]: I0220 08:53:58.398748 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/862b1b6e-55b5-4038-97d3-c1d7bacd7952-catalog-content\") pod \"community-operators-7gx58\" (UID: \"862b1b6e-55b5-4038-97d3-c1d7bacd7952\") " pod="openshift-marketplace/community-operators-7gx58" Feb 20 08:53:58 crc kubenswrapper[4948]: I0220 08:53:58.419288 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtpqf\" (UniqueName: \"kubernetes.io/projected/862b1b6e-55b5-4038-97d3-c1d7bacd7952-kube-api-access-wtpqf\") pod \"community-operators-7gx58\" (UID: \"862b1b6e-55b5-4038-97d3-c1d7bacd7952\") " pod="openshift-marketplace/community-operators-7gx58" Feb 20 08:53:58 crc kubenswrapper[4948]: I0220 08:53:58.566350 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7gx58" Feb 20 08:53:59 crc kubenswrapper[4948]: I0220 08:53:59.032901 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7gx58"] Feb 20 08:53:59 crc kubenswrapper[4948]: I0220 08:53:59.184205 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7gx58" event={"ID":"862b1b6e-55b5-4038-97d3-c1d7bacd7952","Type":"ContainerStarted","Data":"57a392d54b60ec1f9ea44e12e9c7761f58441d678cc91dba5253e1e486cc1739"} Feb 20 08:54:00 crc kubenswrapper[4948]: I0220 08:54:00.201198 4948 generic.go:334] "Generic (PLEG): container finished" podID="862b1b6e-55b5-4038-97d3-c1d7bacd7952" containerID="5cc46dc46f6bf2327c20993dbc63a5a80831fa33a8858c4268efa9310d184072" exitCode=0 Feb 20 08:54:00 crc kubenswrapper[4948]: I0220 08:54:00.201271 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7gx58" event={"ID":"862b1b6e-55b5-4038-97d3-c1d7bacd7952","Type":"ContainerDied","Data":"5cc46dc46f6bf2327c20993dbc63a5a80831fa33a8858c4268efa9310d184072"} Feb 20 08:54:02 crc kubenswrapper[4948]: I0220 08:54:02.516238 4948 generic.go:334] "Generic (PLEG): container finished" podID="862b1b6e-55b5-4038-97d3-c1d7bacd7952" containerID="4d29bdce5261c62a695692c37c480bd775f65ba1c115d881b2d004a4ee0fd7b6" exitCode=0 Feb 20 08:54:02 crc kubenswrapper[4948]: I0220 08:54:02.516764 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7gx58" event={"ID":"862b1b6e-55b5-4038-97d3-c1d7bacd7952","Type":"ContainerDied","Data":"4d29bdce5261c62a695692c37c480bd775f65ba1c115d881b2d004a4ee0fd7b6"} Feb 20 08:54:03 crc kubenswrapper[4948]: I0220 08:54:03.530301 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7gx58" event={"ID":"862b1b6e-55b5-4038-97d3-c1d7bacd7952","Type":"ContainerStarted","Data":"72e21e1986add64697efd8d697ba5a97302e08211fa86c5c48d50e610852427e"} Feb 20 08:54:03 crc kubenswrapper[4948]: I0220 08:54:03.562935 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7gx58" podStartSLOduration=2.8086998359999997 podStartE2EDuration="5.562911118s" podCreationTimestamp="2026-02-20 08:53:58 +0000 UTC" firstStartedPulling="2026-02-20 08:54:00.204362761 +0000 UTC m=+2889.178857581" lastFinishedPulling="2026-02-20 08:54:02.958574013 +0000 UTC m=+2891.933068863" observedRunningTime="2026-02-20 08:54:03.553469047 +0000 UTC m=+2892.527963877" watchObservedRunningTime="2026-02-20 08:54:03.562911118 +0000 UTC m=+2892.537405948" Feb 20 08:54:03 crc kubenswrapper[4948]: I0220 08:54:03.722919 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:54:03 crc kubenswrapper[4948]: E0220 08:54:03.723225 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:54:05 crc kubenswrapper[4948]: I0220 08:54:05.228786 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dqcs5" Feb 20 08:54:05 crc kubenswrapper[4948]: I0220 08:54:05.284187 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dqcs5" Feb 20 08:54:05 crc kubenswrapper[4948]: I0220 08:54:05.460284 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dqcs5"] Feb 20 08:54:06 crc kubenswrapper[4948]: I0220 08:54:06.566189 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dqcs5" podUID="5cb0bef2-1a21-46e7-905d-6dc6d0c50cab" containerName="registry-server" containerID="cri-o://b8438f9e17f81afa0b3fb46a37e2e751ee4f1773769441d7d00fc72148764d6d" gracePeriod=2 Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.097414 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dqcs5" Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.171619 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q77f9\" (UniqueName: \"kubernetes.io/projected/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-kube-api-access-q77f9\") pod \"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab\" (UID: \"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab\") " Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.171757 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-catalog-content\") pod \"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab\" (UID: \"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab\") " Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.172084 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-utilities\") pod \"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab\" (UID: \"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab\") " Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.172760 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-utilities" (OuterVolumeSpecName: "utilities") pod "5cb0bef2-1a21-46e7-905d-6dc6d0c50cab" (UID: "5cb0bef2-1a21-46e7-905d-6dc6d0c50cab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.172904 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.177794 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-kube-api-access-q77f9" (OuterVolumeSpecName: "kube-api-access-q77f9") pod "5cb0bef2-1a21-46e7-905d-6dc6d0c50cab" (UID: "5cb0bef2-1a21-46e7-905d-6dc6d0c50cab"). InnerVolumeSpecName "kube-api-access-q77f9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.275465 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q77f9\" (UniqueName: \"kubernetes.io/projected/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-kube-api-access-q77f9\") on node \"crc\" DevicePath \"\"" Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.285916 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5cb0bef2-1a21-46e7-905d-6dc6d0c50cab" (UID: "5cb0bef2-1a21-46e7-905d-6dc6d0c50cab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.377253 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.576103 4948 generic.go:334] "Generic (PLEG): container finished" podID="5cb0bef2-1a21-46e7-905d-6dc6d0c50cab" containerID="b8438f9e17f81afa0b3fb46a37e2e751ee4f1773769441d7d00fc72148764d6d" exitCode=0 Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.576147 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqcs5" event={"ID":"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab","Type":"ContainerDied","Data":"b8438f9e17f81afa0b3fb46a37e2e751ee4f1773769441d7d00fc72148764d6d"} Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.576169 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dqcs5" Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.576301 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqcs5" event={"ID":"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab","Type":"ContainerDied","Data":"82329664b3d13e9d5f6f72bfbd0c9bfcb4cdedf3613453d9b8308754600e8fce"} Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.576334 4948 scope.go:117] "RemoveContainer" containerID="b8438f9e17f81afa0b3fb46a37e2e751ee4f1773769441d7d00fc72148764d6d" Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.616762 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dqcs5"] Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.619252 4948 scope.go:117] "RemoveContainer" containerID="cdc48d4ad6f6ad6329e706c857c3263754fc1d230af2ba67afcff789d58a0f8f" Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.626787 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dqcs5"] Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.646651 4948 scope.go:117] "RemoveContainer" containerID="e61abe7ff267f9bfeba63a25cfd6a81f5a070c5f25f48e0d6d8f167a049e5620" Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.679028 4948 scope.go:117] "RemoveContainer" containerID="b8438f9e17f81afa0b3fb46a37e2e751ee4f1773769441d7d00fc72148764d6d" Feb 20 08:54:07 crc kubenswrapper[4948]: E0220 08:54:07.679488 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8438f9e17f81afa0b3fb46a37e2e751ee4f1773769441d7d00fc72148764d6d\": container with ID starting with b8438f9e17f81afa0b3fb46a37e2e751ee4f1773769441d7d00fc72148764d6d not found: ID does not exist" containerID="b8438f9e17f81afa0b3fb46a37e2e751ee4f1773769441d7d00fc72148764d6d" Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.679520 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8438f9e17f81afa0b3fb46a37e2e751ee4f1773769441d7d00fc72148764d6d"} err="failed to get container status \"b8438f9e17f81afa0b3fb46a37e2e751ee4f1773769441d7d00fc72148764d6d\": rpc error: code = NotFound desc = could not find container \"b8438f9e17f81afa0b3fb46a37e2e751ee4f1773769441d7d00fc72148764d6d\": container with ID starting with b8438f9e17f81afa0b3fb46a37e2e751ee4f1773769441d7d00fc72148764d6d not found: ID does not exist" Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.679544 4948 scope.go:117] "RemoveContainer" containerID="cdc48d4ad6f6ad6329e706c857c3263754fc1d230af2ba67afcff789d58a0f8f" Feb 20 08:54:07 crc kubenswrapper[4948]: E0220 08:54:07.679888 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdc48d4ad6f6ad6329e706c857c3263754fc1d230af2ba67afcff789d58a0f8f\": container with ID starting with cdc48d4ad6f6ad6329e706c857c3263754fc1d230af2ba67afcff789d58a0f8f not found: ID does not exist" containerID="cdc48d4ad6f6ad6329e706c857c3263754fc1d230af2ba67afcff789d58a0f8f" Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.679941 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdc48d4ad6f6ad6329e706c857c3263754fc1d230af2ba67afcff789d58a0f8f"} err="failed to get container status \"cdc48d4ad6f6ad6329e706c857c3263754fc1d230af2ba67afcff789d58a0f8f\": rpc error: code = NotFound desc = could not find container \"cdc48d4ad6f6ad6329e706c857c3263754fc1d230af2ba67afcff789d58a0f8f\": container with ID starting with cdc48d4ad6f6ad6329e706c857c3263754fc1d230af2ba67afcff789d58a0f8f not found: ID does not exist" Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.679985 4948 scope.go:117] "RemoveContainer" containerID="e61abe7ff267f9bfeba63a25cfd6a81f5a070c5f25f48e0d6d8f167a049e5620" Feb 20 08:54:07 crc kubenswrapper[4948]: E0220 08:54:07.680252 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e61abe7ff267f9bfeba63a25cfd6a81f5a070c5f25f48e0d6d8f167a049e5620\": container with ID starting with e61abe7ff267f9bfeba63a25cfd6a81f5a070c5f25f48e0d6d8f167a049e5620 not found: ID does not exist" containerID="e61abe7ff267f9bfeba63a25cfd6a81f5a070c5f25f48e0d6d8f167a049e5620" Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.680282 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e61abe7ff267f9bfeba63a25cfd6a81f5a070c5f25f48e0d6d8f167a049e5620"} err="failed to get container status \"e61abe7ff267f9bfeba63a25cfd6a81f5a070c5f25f48e0d6d8f167a049e5620\": rpc error: code = NotFound desc = could not find container \"e61abe7ff267f9bfeba63a25cfd6a81f5a070c5f25f48e0d6d8f167a049e5620\": container with ID starting with e61abe7ff267f9bfeba63a25cfd6a81f5a070c5f25f48e0d6d8f167a049e5620 not found: ID does not exist" Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.731794 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cb0bef2-1a21-46e7-905d-6dc6d0c50cab" path="/var/lib/kubelet/pods/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab/volumes" Feb 20 08:54:08 crc kubenswrapper[4948]: I0220 08:54:08.567122 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7gx58" Feb 20 08:54:08 crc kubenswrapper[4948]: I0220 08:54:08.568696 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7gx58" Feb 20 08:54:08 crc kubenswrapper[4948]: I0220 08:54:08.631964 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7gx58" Feb 20 08:54:08 crc kubenswrapper[4948]: I0220 08:54:08.696821 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7gx58" Feb 20 08:54:09 crc kubenswrapper[4948]: I0220 08:54:09.861338 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7gx58"] Feb 20 08:54:10 crc kubenswrapper[4948]: I0220 08:54:10.605698 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7gx58" podUID="862b1b6e-55b5-4038-97d3-c1d7bacd7952" containerName="registry-server" containerID="cri-o://72e21e1986add64697efd8d697ba5a97302e08211fa86c5c48d50e610852427e" gracePeriod=2 Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.118543 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7gx58" Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.249344 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/862b1b6e-55b5-4038-97d3-c1d7bacd7952-catalog-content\") pod \"862b1b6e-55b5-4038-97d3-c1d7bacd7952\" (UID: \"862b1b6e-55b5-4038-97d3-c1d7bacd7952\") " Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.249409 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/862b1b6e-55b5-4038-97d3-c1d7bacd7952-utilities\") pod \"862b1b6e-55b5-4038-97d3-c1d7bacd7952\" (UID: \"862b1b6e-55b5-4038-97d3-c1d7bacd7952\") " Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.249585 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtpqf\" (UniqueName: \"kubernetes.io/projected/862b1b6e-55b5-4038-97d3-c1d7bacd7952-kube-api-access-wtpqf\") pod \"862b1b6e-55b5-4038-97d3-c1d7bacd7952\" (UID: \"862b1b6e-55b5-4038-97d3-c1d7bacd7952\") " Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.250225 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/862b1b6e-55b5-4038-97d3-c1d7bacd7952-utilities" (OuterVolumeSpecName: "utilities") pod "862b1b6e-55b5-4038-97d3-c1d7bacd7952" (UID: "862b1b6e-55b5-4038-97d3-c1d7bacd7952"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.255441 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/862b1b6e-55b5-4038-97d3-c1d7bacd7952-kube-api-access-wtpqf" (OuterVolumeSpecName: "kube-api-access-wtpqf") pod "862b1b6e-55b5-4038-97d3-c1d7bacd7952" (UID: "862b1b6e-55b5-4038-97d3-c1d7bacd7952"). InnerVolumeSpecName "kube-api-access-wtpqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.313366 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/862b1b6e-55b5-4038-97d3-c1d7bacd7952-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "862b1b6e-55b5-4038-97d3-c1d7bacd7952" (UID: "862b1b6e-55b5-4038-97d3-c1d7bacd7952"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.352230 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtpqf\" (UniqueName: \"kubernetes.io/projected/862b1b6e-55b5-4038-97d3-c1d7bacd7952-kube-api-access-wtpqf\") on node \"crc\" DevicePath \"\"" Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.352269 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/862b1b6e-55b5-4038-97d3-c1d7bacd7952-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.352282 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/862b1b6e-55b5-4038-97d3-c1d7bacd7952-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.616961 4948 generic.go:334] "Generic (PLEG): container finished" podID="862b1b6e-55b5-4038-97d3-c1d7bacd7952" containerID="72e21e1986add64697efd8d697ba5a97302e08211fa86c5c48d50e610852427e" exitCode=0 Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.617043 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7gx58" event={"ID":"862b1b6e-55b5-4038-97d3-c1d7bacd7952","Type":"ContainerDied","Data":"72e21e1986add64697efd8d697ba5a97302e08211fa86c5c48d50e610852427e"} Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.617109 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7gx58" Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.617137 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7gx58" event={"ID":"862b1b6e-55b5-4038-97d3-c1d7bacd7952","Type":"ContainerDied","Data":"57a392d54b60ec1f9ea44e12e9c7761f58441d678cc91dba5253e1e486cc1739"} Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.617177 4948 scope.go:117] "RemoveContainer" containerID="72e21e1986add64697efd8d697ba5a97302e08211fa86c5c48d50e610852427e" Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.641296 4948 scope.go:117] "RemoveContainer" containerID="4d29bdce5261c62a695692c37c480bd775f65ba1c115d881b2d004a4ee0fd7b6" Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.663851 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7gx58"] Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.678951 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7gx58"] Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.681207 4948 scope.go:117] "RemoveContainer" containerID="5cc46dc46f6bf2327c20993dbc63a5a80831fa33a8858c4268efa9310d184072" Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.721015 4948 scope.go:117] "RemoveContainer" containerID="72e21e1986add64697efd8d697ba5a97302e08211fa86c5c48d50e610852427e" Feb 20 08:54:11 crc kubenswrapper[4948]: E0220 08:54:11.724266 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72e21e1986add64697efd8d697ba5a97302e08211fa86c5c48d50e610852427e\": container with ID starting with 72e21e1986add64697efd8d697ba5a97302e08211fa86c5c48d50e610852427e not found: ID does not exist" containerID="72e21e1986add64697efd8d697ba5a97302e08211fa86c5c48d50e610852427e" Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.724312 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72e21e1986add64697efd8d697ba5a97302e08211fa86c5c48d50e610852427e"} err="failed to get container status \"72e21e1986add64697efd8d697ba5a97302e08211fa86c5c48d50e610852427e\": rpc error: code = NotFound desc = could not find container \"72e21e1986add64697efd8d697ba5a97302e08211fa86c5c48d50e610852427e\": container with ID starting with 72e21e1986add64697efd8d697ba5a97302e08211fa86c5c48d50e610852427e not found: ID does not exist" Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.724338 4948 scope.go:117] "RemoveContainer" containerID="4d29bdce5261c62a695692c37c480bd775f65ba1c115d881b2d004a4ee0fd7b6" Feb 20 08:54:11 crc kubenswrapper[4948]: E0220 08:54:11.724788 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d29bdce5261c62a695692c37c480bd775f65ba1c115d881b2d004a4ee0fd7b6\": container with ID starting with 4d29bdce5261c62a695692c37c480bd775f65ba1c115d881b2d004a4ee0fd7b6 not found: ID does not exist" containerID="4d29bdce5261c62a695692c37c480bd775f65ba1c115d881b2d004a4ee0fd7b6" Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.724817 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d29bdce5261c62a695692c37c480bd775f65ba1c115d881b2d004a4ee0fd7b6"} err="failed to get container status \"4d29bdce5261c62a695692c37c480bd775f65ba1c115d881b2d004a4ee0fd7b6\": rpc error: code = NotFound desc = could not find container \"4d29bdce5261c62a695692c37c480bd775f65ba1c115d881b2d004a4ee0fd7b6\": container with ID starting with 4d29bdce5261c62a695692c37c480bd775f65ba1c115d881b2d004a4ee0fd7b6 not found: ID does not exist" Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.724832 4948 scope.go:117] "RemoveContainer" containerID="5cc46dc46f6bf2327c20993dbc63a5a80831fa33a8858c4268efa9310d184072" Feb 20 08:54:11 crc kubenswrapper[4948]: E0220 08:54:11.725155 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5cc46dc46f6bf2327c20993dbc63a5a80831fa33a8858c4268efa9310d184072\": container with ID starting with 5cc46dc46f6bf2327c20993dbc63a5a80831fa33a8858c4268efa9310d184072 not found: ID does not exist" containerID="5cc46dc46f6bf2327c20993dbc63a5a80831fa33a8858c4268efa9310d184072" Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.725187 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cc46dc46f6bf2327c20993dbc63a5a80831fa33a8858c4268efa9310d184072"} err="failed to get container status \"5cc46dc46f6bf2327c20993dbc63a5a80831fa33a8858c4268efa9310d184072\": rpc error: code = NotFound desc = could not find container \"5cc46dc46f6bf2327c20993dbc63a5a80831fa33a8858c4268efa9310d184072\": container with ID starting with 5cc46dc46f6bf2327c20993dbc63a5a80831fa33a8858c4268efa9310d184072 not found: ID does not exist" Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.733896 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="862b1b6e-55b5-4038-97d3-c1d7bacd7952" path="/var/lib/kubelet/pods/862b1b6e-55b5-4038-97d3-c1d7bacd7952/volumes" Feb 20 08:54:17 crc kubenswrapper[4948]: I0220 08:54:17.723009 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:54:17 crc kubenswrapper[4948]: E0220 08:54:17.724073 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:54:29 crc kubenswrapper[4948]: I0220 08:54:29.723164 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:54:29 crc kubenswrapper[4948]: E0220 08:54:29.724128 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.183543 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6gxcp"] Feb 20 08:54:36 crc kubenswrapper[4948]: E0220 08:54:36.184639 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="862b1b6e-55b5-4038-97d3-c1d7bacd7952" containerName="extract-utilities" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.184660 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="862b1b6e-55b5-4038-97d3-c1d7bacd7952" containerName="extract-utilities" Feb 20 08:54:36 crc kubenswrapper[4948]: E0220 08:54:36.184673 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cb0bef2-1a21-46e7-905d-6dc6d0c50cab" containerName="extract-content" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.184681 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cb0bef2-1a21-46e7-905d-6dc6d0c50cab" containerName="extract-content" Feb 20 08:54:36 crc kubenswrapper[4948]: E0220 08:54:36.184699 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cb0bef2-1a21-46e7-905d-6dc6d0c50cab" containerName="registry-server" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.184707 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cb0bef2-1a21-46e7-905d-6dc6d0c50cab" containerName="registry-server" Feb 20 08:54:36 crc kubenswrapper[4948]: E0220 08:54:36.184720 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="862b1b6e-55b5-4038-97d3-c1d7bacd7952" containerName="extract-content" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.184727 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="862b1b6e-55b5-4038-97d3-c1d7bacd7952" containerName="extract-content" Feb 20 08:54:36 crc kubenswrapper[4948]: E0220 08:54:36.184756 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cb0bef2-1a21-46e7-905d-6dc6d0c50cab" containerName="extract-utilities" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.184763 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cb0bef2-1a21-46e7-905d-6dc6d0c50cab" containerName="extract-utilities" Feb 20 08:54:36 crc kubenswrapper[4948]: E0220 08:54:36.184774 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="862b1b6e-55b5-4038-97d3-c1d7bacd7952" containerName="registry-server" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.184780 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="862b1b6e-55b5-4038-97d3-c1d7bacd7952" containerName="registry-server" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.185026 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="862b1b6e-55b5-4038-97d3-c1d7bacd7952" containerName="registry-server" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.185054 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cb0bef2-1a21-46e7-905d-6dc6d0c50cab" containerName="registry-server" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.186691 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6gxcp" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.210021 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6gxcp"] Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.247061 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddsfc\" (UniqueName: \"kubernetes.io/projected/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-kube-api-access-ddsfc\") pod \"certified-operators-6gxcp\" (UID: \"0f0a4e55-97e8-4b22-8299-ea443d51a3c1\") " pod="openshift-marketplace/certified-operators-6gxcp" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.248132 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-utilities\") pod \"certified-operators-6gxcp\" (UID: \"0f0a4e55-97e8-4b22-8299-ea443d51a3c1\") " pod="openshift-marketplace/certified-operators-6gxcp" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.248234 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-catalog-content\") pod \"certified-operators-6gxcp\" (UID: \"0f0a4e55-97e8-4b22-8299-ea443d51a3c1\") " pod="openshift-marketplace/certified-operators-6gxcp" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.350145 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddsfc\" (UniqueName: \"kubernetes.io/projected/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-kube-api-access-ddsfc\") pod \"certified-operators-6gxcp\" (UID: \"0f0a4e55-97e8-4b22-8299-ea443d51a3c1\") " pod="openshift-marketplace/certified-operators-6gxcp" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.350239 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-utilities\") pod \"certified-operators-6gxcp\" (UID: \"0f0a4e55-97e8-4b22-8299-ea443d51a3c1\") " pod="openshift-marketplace/certified-operators-6gxcp" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.350289 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-catalog-content\") pod \"certified-operators-6gxcp\" (UID: \"0f0a4e55-97e8-4b22-8299-ea443d51a3c1\") " pod="openshift-marketplace/certified-operators-6gxcp" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.351006 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-catalog-content\") pod \"certified-operators-6gxcp\" (UID: \"0f0a4e55-97e8-4b22-8299-ea443d51a3c1\") " pod="openshift-marketplace/certified-operators-6gxcp" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.351029 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-utilities\") pod \"certified-operators-6gxcp\" (UID: \"0f0a4e55-97e8-4b22-8299-ea443d51a3c1\") " pod="openshift-marketplace/certified-operators-6gxcp" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.371884 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddsfc\" (UniqueName: \"kubernetes.io/projected/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-kube-api-access-ddsfc\") pod \"certified-operators-6gxcp\" (UID: \"0f0a4e55-97e8-4b22-8299-ea443d51a3c1\") " pod="openshift-marketplace/certified-operators-6gxcp" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.515659 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6gxcp" Feb 20 08:54:37 crc kubenswrapper[4948]: I0220 08:54:37.083095 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6gxcp"] Feb 20 08:54:37 crc kubenswrapper[4948]: I0220 08:54:37.887048 4948 generic.go:334] "Generic (PLEG): container finished" podID="0f0a4e55-97e8-4b22-8299-ea443d51a3c1" containerID="d0716005a1627047c505cf076f1e700308f61ddebea1580bef3289c1f5855a73" exitCode=0 Feb 20 08:54:37 crc kubenswrapper[4948]: I0220 08:54:37.887115 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6gxcp" event={"ID":"0f0a4e55-97e8-4b22-8299-ea443d51a3c1","Type":"ContainerDied","Data":"d0716005a1627047c505cf076f1e700308f61ddebea1580bef3289c1f5855a73"} Feb 20 08:54:37 crc kubenswrapper[4948]: I0220 08:54:37.887181 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6gxcp" event={"ID":"0f0a4e55-97e8-4b22-8299-ea443d51a3c1","Type":"ContainerStarted","Data":"653f289bbb6954dacf1dccca73b66f05c7ef3ea91b0fb94146d797c22d4827b2"} Feb 20 08:54:38 crc kubenswrapper[4948]: I0220 08:54:38.899375 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6gxcp" event={"ID":"0f0a4e55-97e8-4b22-8299-ea443d51a3c1","Type":"ContainerStarted","Data":"1028c59d0521f3ece4f33a9c808df88e3f3aeb440639acaf3addacfc8e3fc9f3"} Feb 20 08:54:39 crc kubenswrapper[4948]: I0220 08:54:39.911966 4948 generic.go:334] "Generic (PLEG): container finished" podID="0f0a4e55-97e8-4b22-8299-ea443d51a3c1" containerID="1028c59d0521f3ece4f33a9c808df88e3f3aeb440639acaf3addacfc8e3fc9f3" exitCode=0 Feb 20 08:54:39 crc kubenswrapper[4948]: I0220 08:54:39.912032 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6gxcp" event={"ID":"0f0a4e55-97e8-4b22-8299-ea443d51a3c1","Type":"ContainerDied","Data":"1028c59d0521f3ece4f33a9c808df88e3f3aeb440639acaf3addacfc8e3fc9f3"} Feb 20 08:54:40 crc kubenswrapper[4948]: I0220 08:54:40.723096 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:54:40 crc kubenswrapper[4948]: E0220 08:54:40.723902 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:54:40 crc kubenswrapper[4948]: I0220 08:54:40.925475 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6gxcp" event={"ID":"0f0a4e55-97e8-4b22-8299-ea443d51a3c1","Type":"ContainerStarted","Data":"1ae5da12074875624d809ff770ce45f4eed832a8a73b4e198a873dfd6f9b6b34"} Feb 20 08:54:40 crc kubenswrapper[4948]: I0220 08:54:40.949672 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6gxcp" podStartSLOduration=2.526167647 podStartE2EDuration="4.949644825s" podCreationTimestamp="2026-02-20 08:54:36 +0000 UTC" firstStartedPulling="2026-02-20 08:54:37.889425737 +0000 UTC m=+2926.863920557" lastFinishedPulling="2026-02-20 08:54:40.312902915 +0000 UTC m=+2929.287397735" observedRunningTime="2026-02-20 08:54:40.943043453 +0000 UTC m=+2929.917538273" watchObservedRunningTime="2026-02-20 08:54:40.949644825 +0000 UTC m=+2929.924139645" Feb 20 08:54:46 crc kubenswrapper[4948]: I0220 08:54:46.517037 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6gxcp" Feb 20 08:54:46 crc kubenswrapper[4948]: I0220 08:54:46.517700 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6gxcp" Feb 20 08:54:46 crc kubenswrapper[4948]: I0220 08:54:46.558112 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6gxcp" Feb 20 08:54:47 crc kubenswrapper[4948]: I0220 08:54:47.042722 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6gxcp" Feb 20 08:54:47 crc kubenswrapper[4948]: I0220 08:54:47.105337 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6gxcp"] Feb 20 08:54:49 crc kubenswrapper[4948]: I0220 08:54:49.004525 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6gxcp" podUID="0f0a4e55-97e8-4b22-8299-ea443d51a3c1" containerName="registry-server" containerID="cri-o://1ae5da12074875624d809ff770ce45f4eed832a8a73b4e198a873dfd6f9b6b34" gracePeriod=2 Feb 20 08:54:49 crc kubenswrapper[4948]: I0220 08:54:49.571205 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6gxcp" Feb 20 08:54:49 crc kubenswrapper[4948]: I0220 08:54:49.639290 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-catalog-content\") pod \"0f0a4e55-97e8-4b22-8299-ea443d51a3c1\" (UID: \"0f0a4e55-97e8-4b22-8299-ea443d51a3c1\") " Feb 20 08:54:49 crc kubenswrapper[4948]: I0220 08:54:49.639511 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddsfc\" (UniqueName: \"kubernetes.io/projected/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-kube-api-access-ddsfc\") pod \"0f0a4e55-97e8-4b22-8299-ea443d51a3c1\" (UID: \"0f0a4e55-97e8-4b22-8299-ea443d51a3c1\") " Feb 20 08:54:49 crc kubenswrapper[4948]: I0220 08:54:49.639628 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-utilities\") pod \"0f0a4e55-97e8-4b22-8299-ea443d51a3c1\" (UID: \"0f0a4e55-97e8-4b22-8299-ea443d51a3c1\") " Feb 20 08:54:49 crc kubenswrapper[4948]: I0220 08:54:49.642055 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-utilities" (OuterVolumeSpecName: "utilities") pod "0f0a4e55-97e8-4b22-8299-ea443d51a3c1" (UID: "0f0a4e55-97e8-4b22-8299-ea443d51a3c1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:54:49 crc kubenswrapper[4948]: I0220 08:54:49.652187 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-kube-api-access-ddsfc" (OuterVolumeSpecName: "kube-api-access-ddsfc") pod "0f0a4e55-97e8-4b22-8299-ea443d51a3c1" (UID: "0f0a4e55-97e8-4b22-8299-ea443d51a3c1"). InnerVolumeSpecName "kube-api-access-ddsfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:54:49 crc kubenswrapper[4948]: I0220 08:54:49.748462 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddsfc\" (UniqueName: \"kubernetes.io/projected/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-kube-api-access-ddsfc\") on node \"crc\" DevicePath \"\"" Feb 20 08:54:49 crc kubenswrapper[4948]: I0220 08:54:49.748815 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:54:49 crc kubenswrapper[4948]: I0220 08:54:49.795112 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0f0a4e55-97e8-4b22-8299-ea443d51a3c1" (UID: "0f0a4e55-97e8-4b22-8299-ea443d51a3c1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:54:49 crc kubenswrapper[4948]: I0220 08:54:49.851617 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:54:50 crc kubenswrapper[4948]: I0220 08:54:50.016998 4948 generic.go:334] "Generic (PLEG): container finished" podID="0f0a4e55-97e8-4b22-8299-ea443d51a3c1" containerID="1ae5da12074875624d809ff770ce45f4eed832a8a73b4e198a873dfd6f9b6b34" exitCode=0 Feb 20 08:54:50 crc kubenswrapper[4948]: I0220 08:54:50.017058 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6gxcp" event={"ID":"0f0a4e55-97e8-4b22-8299-ea443d51a3c1","Type":"ContainerDied","Data":"1ae5da12074875624d809ff770ce45f4eed832a8a73b4e198a873dfd6f9b6b34"} Feb 20 08:54:50 crc kubenswrapper[4948]: I0220 08:54:50.017095 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6gxcp" event={"ID":"0f0a4e55-97e8-4b22-8299-ea443d51a3c1","Type":"ContainerDied","Data":"653f289bbb6954dacf1dccca73b66f05c7ef3ea91b0fb94146d797c22d4827b2"} Feb 20 08:54:50 crc kubenswrapper[4948]: I0220 08:54:50.017117 4948 scope.go:117] "RemoveContainer" containerID="1ae5da12074875624d809ff770ce45f4eed832a8a73b4e198a873dfd6f9b6b34" Feb 20 08:54:50 crc kubenswrapper[4948]: I0220 08:54:50.017296 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6gxcp" Feb 20 08:54:50 crc kubenswrapper[4948]: I0220 08:54:50.040679 4948 scope.go:117] "RemoveContainer" containerID="1028c59d0521f3ece4f33a9c808df88e3f3aeb440639acaf3addacfc8e3fc9f3" Feb 20 08:54:50 crc kubenswrapper[4948]: I0220 08:54:50.083169 4948 scope.go:117] "RemoveContainer" containerID="d0716005a1627047c505cf076f1e700308f61ddebea1580bef3289c1f5855a73" Feb 20 08:54:50 crc kubenswrapper[4948]: I0220 08:54:50.085953 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6gxcp"] Feb 20 08:54:50 crc kubenswrapper[4948]: I0220 08:54:50.097402 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6gxcp"] Feb 20 08:54:50 crc kubenswrapper[4948]: I0220 08:54:50.133836 4948 scope.go:117] "RemoveContainer" containerID="1ae5da12074875624d809ff770ce45f4eed832a8a73b4e198a873dfd6f9b6b34" Feb 20 08:54:50 crc kubenswrapper[4948]: E0220 08:54:50.134449 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ae5da12074875624d809ff770ce45f4eed832a8a73b4e198a873dfd6f9b6b34\": container with ID starting with 1ae5da12074875624d809ff770ce45f4eed832a8a73b4e198a873dfd6f9b6b34 not found: ID does not exist" containerID="1ae5da12074875624d809ff770ce45f4eed832a8a73b4e198a873dfd6f9b6b34" Feb 20 08:54:50 crc kubenswrapper[4948]: I0220 08:54:50.134482 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ae5da12074875624d809ff770ce45f4eed832a8a73b4e198a873dfd6f9b6b34"} err="failed to get container status \"1ae5da12074875624d809ff770ce45f4eed832a8a73b4e198a873dfd6f9b6b34\": rpc error: code = NotFound desc = could not find container \"1ae5da12074875624d809ff770ce45f4eed832a8a73b4e198a873dfd6f9b6b34\": container with ID starting with 1ae5da12074875624d809ff770ce45f4eed832a8a73b4e198a873dfd6f9b6b34 not found: ID does not exist" Feb 20 08:54:50 crc kubenswrapper[4948]: I0220 08:54:50.134507 4948 scope.go:117] "RemoveContainer" containerID="1028c59d0521f3ece4f33a9c808df88e3f3aeb440639acaf3addacfc8e3fc9f3" Feb 20 08:54:50 crc kubenswrapper[4948]: E0220 08:54:50.135556 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1028c59d0521f3ece4f33a9c808df88e3f3aeb440639acaf3addacfc8e3fc9f3\": container with ID starting with 1028c59d0521f3ece4f33a9c808df88e3f3aeb440639acaf3addacfc8e3fc9f3 not found: ID does not exist" containerID="1028c59d0521f3ece4f33a9c808df88e3f3aeb440639acaf3addacfc8e3fc9f3" Feb 20 08:54:50 crc kubenswrapper[4948]: I0220 08:54:50.135593 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1028c59d0521f3ece4f33a9c808df88e3f3aeb440639acaf3addacfc8e3fc9f3"} err="failed to get container status \"1028c59d0521f3ece4f33a9c808df88e3f3aeb440639acaf3addacfc8e3fc9f3\": rpc error: code = NotFound desc = could not find container \"1028c59d0521f3ece4f33a9c808df88e3f3aeb440639acaf3addacfc8e3fc9f3\": container with ID starting with 1028c59d0521f3ece4f33a9c808df88e3f3aeb440639acaf3addacfc8e3fc9f3 not found: ID does not exist" Feb 20 08:54:50 crc kubenswrapper[4948]: I0220 08:54:50.135612 4948 scope.go:117] "RemoveContainer" containerID="d0716005a1627047c505cf076f1e700308f61ddebea1580bef3289c1f5855a73" Feb 20 08:54:50 crc kubenswrapper[4948]: E0220 08:54:50.135871 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0716005a1627047c505cf076f1e700308f61ddebea1580bef3289c1f5855a73\": container with ID starting with d0716005a1627047c505cf076f1e700308f61ddebea1580bef3289c1f5855a73 not found: ID does not exist" containerID="d0716005a1627047c505cf076f1e700308f61ddebea1580bef3289c1f5855a73" Feb 20 08:54:50 crc kubenswrapper[4948]: I0220 08:54:50.135897 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0716005a1627047c505cf076f1e700308f61ddebea1580bef3289c1f5855a73"} err="failed to get container status \"d0716005a1627047c505cf076f1e700308f61ddebea1580bef3289c1f5855a73\": rpc error: code = NotFound desc = could not find container \"d0716005a1627047c505cf076f1e700308f61ddebea1580bef3289c1f5855a73\": container with ID starting with d0716005a1627047c505cf076f1e700308f61ddebea1580bef3289c1f5855a73 not found: ID does not exist" Feb 20 08:54:51 crc kubenswrapper[4948]: I0220 08:54:51.741830 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f0a4e55-97e8-4b22-8299-ea443d51a3c1" path="/var/lib/kubelet/pods/0f0a4e55-97e8-4b22-8299-ea443d51a3c1/volumes" Feb 20 08:54:54 crc kubenswrapper[4948]: I0220 08:54:54.723153 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:54:54 crc kubenswrapper[4948]: E0220 08:54:54.724166 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:55:09 crc kubenswrapper[4948]: I0220 08:55:09.723150 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:55:09 crc kubenswrapper[4948]: E0220 08:55:09.724016 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:55:20 crc kubenswrapper[4948]: I0220 08:55:20.723367 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:55:20 crc kubenswrapper[4948]: E0220 08:55:20.724796 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:55:33 crc kubenswrapper[4948]: I0220 08:55:33.724643 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:55:33 crc kubenswrapper[4948]: E0220 08:55:33.725894 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:55:47 crc kubenswrapper[4948]: I0220 08:55:47.747193 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:55:47 crc kubenswrapper[4948]: E0220 08:55:47.749372 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:56:00 crc kubenswrapper[4948]: I0220 08:56:00.722957 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:56:00 crc kubenswrapper[4948]: E0220 08:56:00.723749 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:56:11 crc kubenswrapper[4948]: I0220 08:56:11.728796 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:56:11 crc kubenswrapper[4948]: E0220 08:56:11.730052 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:56:26 crc kubenswrapper[4948]: I0220 08:56:26.724214 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:56:26 crc kubenswrapper[4948]: E0220 08:56:26.725765 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:56:41 crc kubenswrapper[4948]: I0220 08:56:41.736878 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:56:41 crc kubenswrapper[4948]: E0220 08:56:41.738370 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:56:53 crc kubenswrapper[4948]: I0220 08:56:53.722749 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:56:53 crc kubenswrapper[4948]: E0220 08:56:53.724004 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:57:05 crc kubenswrapper[4948]: I0220 08:57:05.723232 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:57:05 crc kubenswrapper[4948]: E0220 08:57:05.724391 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:57:17 crc kubenswrapper[4948]: I0220 08:57:17.723495 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:57:17 crc kubenswrapper[4948]: E0220 08:57:17.727215 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:57:31 crc kubenswrapper[4948]: I0220 08:57:31.729696 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:57:31 crc kubenswrapper[4948]: E0220 08:57:31.730486 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:57:42 crc kubenswrapper[4948]: I0220 08:57:42.722312 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:57:42 crc kubenswrapper[4948]: E0220 08:57:42.723112 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:57:56 crc kubenswrapper[4948]: I0220 08:57:56.722942 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:57:56 crc kubenswrapper[4948]: E0220 08:57:56.723828 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:58:10 crc kubenswrapper[4948]: I0220 08:58:10.723145 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:58:10 crc kubenswrapper[4948]: E0220 08:58:10.724575 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:58:24 crc kubenswrapper[4948]: I0220 08:58:24.722239 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:58:24 crc kubenswrapper[4948]: E0220 08:58:24.723108 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:58:36 crc kubenswrapper[4948]: I0220 08:58:36.723154 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:58:36 crc kubenswrapper[4948]: E0220 08:58:36.725801 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:58:49 crc kubenswrapper[4948]: I0220 08:58:49.723754 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:58:50 crc kubenswrapper[4948]: I0220 08:58:50.521317 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"b0e4da4a6a513a5c47c1f0dd87dcf0b56d3fede7e9008d839ac5623f61075ff8"} Feb 20 08:59:07 crc kubenswrapper[4948]: I0220 08:59:07.759124 4948 generic.go:334] "Generic (PLEG): container finished" podID="0fefa0c4-ece9-4caf-8b41-8a64472c53e6" containerID="7c1a082fb78288f20cddeb8b89d13f95d3e809472fa926ab7427ecd860311a72" exitCode=0 Feb 20 08:59:07 crc kubenswrapper[4948]: I0220 08:59:07.759268 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"0fefa0c4-ece9-4caf-8b41-8a64472c53e6","Type":"ContainerDied","Data":"7c1a082fb78288f20cddeb8b89d13f95d3e809472fa926ab7427ecd860311a72"} Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.245400 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.361597 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.361707 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-openstack-config\") pod \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.361780 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-openstack-config-secret\") pod \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.361881 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dtx4h\" (UniqueName: \"kubernetes.io/projected/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-kube-api-access-dtx4h\") pod \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.361946 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-config-data\") pod \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.362063 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-ca-certs\") pod \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.362210 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-ssh-key\") pod \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.362450 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-test-operator-ephemeral-temporary\") pod \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.362543 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-test-operator-ephemeral-workdir\") pod \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.362905 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "0fefa0c4-ece9-4caf-8b41-8a64472c53e6" (UID: "0fefa0c4-ece9-4caf-8b41-8a64472c53e6"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.362706 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-config-data" (OuterVolumeSpecName: "config-data") pod "0fefa0c4-ece9-4caf-8b41-8a64472c53e6" (UID: "0fefa0c4-ece9-4caf-8b41-8a64472c53e6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.363508 4948 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.363551 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.367420 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "0fefa0c4-ece9-4caf-8b41-8a64472c53e6" (UID: "0fefa0c4-ece9-4caf-8b41-8a64472c53e6"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.368007 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "test-operator-logs") pod "0fefa0c4-ece9-4caf-8b41-8a64472c53e6" (UID: "0fefa0c4-ece9-4caf-8b41-8a64472c53e6"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.368137 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-kube-api-access-dtx4h" (OuterVolumeSpecName: "kube-api-access-dtx4h") pod "0fefa0c4-ece9-4caf-8b41-8a64472c53e6" (UID: "0fefa0c4-ece9-4caf-8b41-8a64472c53e6"). InnerVolumeSpecName "kube-api-access-dtx4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.393336 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "0fefa0c4-ece9-4caf-8b41-8a64472c53e6" (UID: "0fefa0c4-ece9-4caf-8b41-8a64472c53e6"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.396057 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0fefa0c4-ece9-4caf-8b41-8a64472c53e6" (UID: "0fefa0c4-ece9-4caf-8b41-8a64472c53e6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.402339 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "0fefa0c4-ece9-4caf-8b41-8a64472c53e6" (UID: "0fefa0c4-ece9-4caf-8b41-8a64472c53e6"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.414556 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "0fefa0c4-ece9-4caf-8b41-8a64472c53e6" (UID: "0fefa0c4-ece9-4caf-8b41-8a64472c53e6"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.465798 4948 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-openstack-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.465855 4948 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.465876 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dtx4h\" (UniqueName: \"kubernetes.io/projected/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-kube-api-access-dtx4h\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.465894 4948 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-ca-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.465910 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-ssh-key\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.465941 4948 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.466075 4948 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.512043 4948 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.567879 4948 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.785626 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"0fefa0c4-ece9-4caf-8b41-8a64472c53e6","Type":"ContainerDied","Data":"8d7b11cefa454a0dc17ba4434816118eb89ddebb724fafd4bee4a5317c66243c"} Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.785671 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d7b11cefa454a0dc17ba4434816118eb89ddebb724fafd4bee4a5317c66243c" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.785771 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.378250 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-g72b5"] Feb 20 08:59:11 crc kubenswrapper[4948]: E0220 08:59:11.379069 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f0a4e55-97e8-4b22-8299-ea443d51a3c1" containerName="extract-utilities" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.379086 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f0a4e55-97e8-4b22-8299-ea443d51a3c1" containerName="extract-utilities" Feb 20 08:59:11 crc kubenswrapper[4948]: E0220 08:59:11.379112 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f0a4e55-97e8-4b22-8299-ea443d51a3c1" containerName="extract-content" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.379121 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f0a4e55-97e8-4b22-8299-ea443d51a3c1" containerName="extract-content" Feb 20 08:59:11 crc kubenswrapper[4948]: E0220 08:59:11.379145 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f0a4e55-97e8-4b22-8299-ea443d51a3c1" containerName="registry-server" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.379153 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f0a4e55-97e8-4b22-8299-ea443d51a3c1" containerName="registry-server" Feb 20 08:59:11 crc kubenswrapper[4948]: E0220 08:59:11.379177 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fefa0c4-ece9-4caf-8b41-8a64472c53e6" containerName="tempest-tests-tempest-tests-runner" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.379186 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fefa0c4-ece9-4caf-8b41-8a64472c53e6" containerName="tempest-tests-tempest-tests-runner" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.379458 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f0a4e55-97e8-4b22-8299-ea443d51a3c1" containerName="registry-server" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.379477 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fefa0c4-ece9-4caf-8b41-8a64472c53e6" containerName="tempest-tests-tempest-tests-runner" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.395120 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g72b5" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.398224 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g72b5"] Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.523090 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c730f22-7ac4-4c78-b897-5aed76dbbd79-utilities\") pod \"redhat-marketplace-g72b5\" (UID: \"4c730f22-7ac4-4c78-b897-5aed76dbbd79\") " pod="openshift-marketplace/redhat-marketplace-g72b5" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.523149 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srqhw\" (UniqueName: \"kubernetes.io/projected/4c730f22-7ac4-4c78-b897-5aed76dbbd79-kube-api-access-srqhw\") pod \"redhat-marketplace-g72b5\" (UID: \"4c730f22-7ac4-4c78-b897-5aed76dbbd79\") " pod="openshift-marketplace/redhat-marketplace-g72b5" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.523220 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c730f22-7ac4-4c78-b897-5aed76dbbd79-catalog-content\") pod \"redhat-marketplace-g72b5\" (UID: \"4c730f22-7ac4-4c78-b897-5aed76dbbd79\") " pod="openshift-marketplace/redhat-marketplace-g72b5" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.624753 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c730f22-7ac4-4c78-b897-5aed76dbbd79-utilities\") pod \"redhat-marketplace-g72b5\" (UID: \"4c730f22-7ac4-4c78-b897-5aed76dbbd79\") " pod="openshift-marketplace/redhat-marketplace-g72b5" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.624800 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srqhw\" (UniqueName: \"kubernetes.io/projected/4c730f22-7ac4-4c78-b897-5aed76dbbd79-kube-api-access-srqhw\") pod \"redhat-marketplace-g72b5\" (UID: \"4c730f22-7ac4-4c78-b897-5aed76dbbd79\") " pod="openshift-marketplace/redhat-marketplace-g72b5" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.624839 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c730f22-7ac4-4c78-b897-5aed76dbbd79-catalog-content\") pod \"redhat-marketplace-g72b5\" (UID: \"4c730f22-7ac4-4c78-b897-5aed76dbbd79\") " pod="openshift-marketplace/redhat-marketplace-g72b5" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.625512 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c730f22-7ac4-4c78-b897-5aed76dbbd79-catalog-content\") pod \"redhat-marketplace-g72b5\" (UID: \"4c730f22-7ac4-4c78-b897-5aed76dbbd79\") " pod="openshift-marketplace/redhat-marketplace-g72b5" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.625512 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c730f22-7ac4-4c78-b897-5aed76dbbd79-utilities\") pod \"redhat-marketplace-g72b5\" (UID: \"4c730f22-7ac4-4c78-b897-5aed76dbbd79\") " pod="openshift-marketplace/redhat-marketplace-g72b5" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.643998 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srqhw\" (UniqueName: \"kubernetes.io/projected/4c730f22-7ac4-4c78-b897-5aed76dbbd79-kube-api-access-srqhw\") pod \"redhat-marketplace-g72b5\" (UID: \"4c730f22-7ac4-4c78-b897-5aed76dbbd79\") " pod="openshift-marketplace/redhat-marketplace-g72b5" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.726255 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g72b5" Feb 20 08:59:12 crc kubenswrapper[4948]: I0220 08:59:12.242689 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g72b5"] Feb 20 08:59:12 crc kubenswrapper[4948]: W0220 08:59:12.252361 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c730f22_7ac4_4c78_b897_5aed76dbbd79.slice/crio-834d32f4c058854127072ff368549105e1f01138af7f056fc671dc63c4a176c7 WatchSource:0}: Error finding container 834d32f4c058854127072ff368549105e1f01138af7f056fc671dc63c4a176c7: Status 404 returned error can't find the container with id 834d32f4c058854127072ff368549105e1f01138af7f056fc671dc63c4a176c7 Feb 20 08:59:12 crc kubenswrapper[4948]: I0220 08:59:12.822889 4948 generic.go:334] "Generic (PLEG): container finished" podID="4c730f22-7ac4-4c78-b897-5aed76dbbd79" containerID="f4c2df2bcbe2ce8c1b34e4d9226525be076033ebb476794a5b25b1871ed87821" exitCode=0 Feb 20 08:59:12 crc kubenswrapper[4948]: I0220 08:59:12.822952 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g72b5" event={"ID":"4c730f22-7ac4-4c78-b897-5aed76dbbd79","Type":"ContainerDied","Data":"f4c2df2bcbe2ce8c1b34e4d9226525be076033ebb476794a5b25b1871ed87821"} Feb 20 08:59:12 crc kubenswrapper[4948]: I0220 08:59:12.823027 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g72b5" event={"ID":"4c730f22-7ac4-4c78-b897-5aed76dbbd79","Type":"ContainerStarted","Data":"834d32f4c058854127072ff368549105e1f01138af7f056fc671dc63c4a176c7"} Feb 20 08:59:12 crc kubenswrapper[4948]: I0220 08:59:12.824832 4948 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 08:59:13 crc kubenswrapper[4948]: I0220 08:59:13.836798 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g72b5" event={"ID":"4c730f22-7ac4-4c78-b897-5aed76dbbd79","Type":"ContainerStarted","Data":"f990f9b78c70c0197487e4632df495d206960a5cc7ff502f7b84e44e7c850bf0"} Feb 20 08:59:13 crc kubenswrapper[4948]: I0220 08:59:13.958444 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 20 08:59:13 crc kubenswrapper[4948]: I0220 08:59:13.960082 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 20 08:59:13 crc kubenswrapper[4948]: I0220 08:59:13.962647 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-tbg4r" Feb 20 08:59:13 crc kubenswrapper[4948]: I0220 08:59:13.981186 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 20 08:59:14 crc kubenswrapper[4948]: I0220 08:59:14.070252 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"de083766-efe0-42cd-95c5-b87bbeb33b0a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 20 08:59:14 crc kubenswrapper[4948]: I0220 08:59:14.070564 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9drzn\" (UniqueName: \"kubernetes.io/projected/de083766-efe0-42cd-95c5-b87bbeb33b0a-kube-api-access-9drzn\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"de083766-efe0-42cd-95c5-b87bbeb33b0a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 20 08:59:14 crc kubenswrapper[4948]: I0220 08:59:14.172332 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"de083766-efe0-42cd-95c5-b87bbeb33b0a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 20 08:59:14 crc kubenswrapper[4948]: I0220 08:59:14.172693 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9drzn\" (UniqueName: \"kubernetes.io/projected/de083766-efe0-42cd-95c5-b87bbeb33b0a-kube-api-access-9drzn\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"de083766-efe0-42cd-95c5-b87bbeb33b0a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 20 08:59:14 crc kubenswrapper[4948]: I0220 08:59:14.172934 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"de083766-efe0-42cd-95c5-b87bbeb33b0a\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 20 08:59:14 crc kubenswrapper[4948]: I0220 08:59:14.200998 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9drzn\" (UniqueName: \"kubernetes.io/projected/de083766-efe0-42cd-95c5-b87bbeb33b0a-kube-api-access-9drzn\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"de083766-efe0-42cd-95c5-b87bbeb33b0a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 20 08:59:14 crc kubenswrapper[4948]: I0220 08:59:14.219313 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"de083766-efe0-42cd-95c5-b87bbeb33b0a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 20 08:59:14 crc kubenswrapper[4948]: I0220 08:59:14.289141 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 20 08:59:14 crc kubenswrapper[4948]: I0220 08:59:14.820308 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 20 08:59:14 crc kubenswrapper[4948]: I0220 08:59:14.846477 4948 generic.go:334] "Generic (PLEG): container finished" podID="4c730f22-7ac4-4c78-b897-5aed76dbbd79" containerID="f990f9b78c70c0197487e4632df495d206960a5cc7ff502f7b84e44e7c850bf0" exitCode=0 Feb 20 08:59:14 crc kubenswrapper[4948]: I0220 08:59:14.847450 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g72b5" event={"ID":"4c730f22-7ac4-4c78-b897-5aed76dbbd79","Type":"ContainerDied","Data":"f990f9b78c70c0197487e4632df495d206960a5cc7ff502f7b84e44e7c850bf0"} Feb 20 08:59:14 crc kubenswrapper[4948]: I0220 08:59:14.852003 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"de083766-efe0-42cd-95c5-b87bbeb33b0a","Type":"ContainerStarted","Data":"660b8d90d795d12fd9b82a49aa9566597ff1b439beab84abbc27859e9de18cb7"} Feb 20 08:59:16 crc kubenswrapper[4948]: I0220 08:59:16.868881 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g72b5" event={"ID":"4c730f22-7ac4-4c78-b897-5aed76dbbd79","Type":"ContainerStarted","Data":"144c046a74ee5454148fd8c9d93123fe6564ef046b5f98222707f50a14a587be"} Feb 20 08:59:16 crc kubenswrapper[4948]: I0220 08:59:16.906176 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-g72b5" podStartSLOduration=2.804067953 podStartE2EDuration="5.906155814s" podCreationTimestamp="2026-02-20 08:59:11 +0000 UTC" firstStartedPulling="2026-02-20 08:59:12.824563728 +0000 UTC m=+3201.799058548" lastFinishedPulling="2026-02-20 08:59:15.926651559 +0000 UTC m=+3204.901146409" observedRunningTime="2026-02-20 08:59:16.891947213 +0000 UTC m=+3205.866442033" watchObservedRunningTime="2026-02-20 08:59:16.906155814 +0000 UTC m=+3205.880650634" Feb 20 08:59:17 crc kubenswrapper[4948]: I0220 08:59:17.881627 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"de083766-efe0-42cd-95c5-b87bbeb33b0a","Type":"ContainerStarted","Data":"4b8d29446ac83847d90e3bb8e9583f0225f10b403cd18e4c1734d1f7b374eff9"} Feb 20 08:59:17 crc kubenswrapper[4948]: I0220 08:59:17.905304 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=3.044024304 podStartE2EDuration="4.905269482s" podCreationTimestamp="2026-02-20 08:59:13 +0000 UTC" firstStartedPulling="2026-02-20 08:59:14.81243052 +0000 UTC m=+3203.786925350" lastFinishedPulling="2026-02-20 08:59:16.673675688 +0000 UTC m=+3205.648170528" observedRunningTime="2026-02-20 08:59:17.903003766 +0000 UTC m=+3206.877498626" watchObservedRunningTime="2026-02-20 08:59:17.905269482 +0000 UTC m=+3206.879764342" Feb 20 08:59:21 crc kubenswrapper[4948]: I0220 08:59:21.742262 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-g72b5" Feb 20 08:59:21 crc kubenswrapper[4948]: I0220 08:59:21.742818 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-g72b5" Feb 20 08:59:21 crc kubenswrapper[4948]: I0220 08:59:21.792772 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-g72b5" Feb 20 08:59:21 crc kubenswrapper[4948]: I0220 08:59:21.973786 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-g72b5" Feb 20 08:59:22 crc kubenswrapper[4948]: I0220 08:59:22.048277 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-g72b5"] Feb 20 08:59:23 crc kubenswrapper[4948]: I0220 08:59:23.949654 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-g72b5" podUID="4c730f22-7ac4-4c78-b897-5aed76dbbd79" containerName="registry-server" containerID="cri-o://144c046a74ee5454148fd8c9d93123fe6564ef046b5f98222707f50a14a587be" gracePeriod=2 Feb 20 08:59:24 crc kubenswrapper[4948]: I0220 08:59:24.453721 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g72b5" Feb 20 08:59:24 crc kubenswrapper[4948]: I0220 08:59:24.633850 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c730f22-7ac4-4c78-b897-5aed76dbbd79-catalog-content\") pod \"4c730f22-7ac4-4c78-b897-5aed76dbbd79\" (UID: \"4c730f22-7ac4-4c78-b897-5aed76dbbd79\") " Feb 20 08:59:24 crc kubenswrapper[4948]: I0220 08:59:24.634146 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srqhw\" (UniqueName: \"kubernetes.io/projected/4c730f22-7ac4-4c78-b897-5aed76dbbd79-kube-api-access-srqhw\") pod \"4c730f22-7ac4-4c78-b897-5aed76dbbd79\" (UID: \"4c730f22-7ac4-4c78-b897-5aed76dbbd79\") " Feb 20 08:59:24 crc kubenswrapper[4948]: I0220 08:59:24.634198 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c730f22-7ac4-4c78-b897-5aed76dbbd79-utilities\") pod \"4c730f22-7ac4-4c78-b897-5aed76dbbd79\" (UID: \"4c730f22-7ac4-4c78-b897-5aed76dbbd79\") " Feb 20 08:59:24 crc kubenswrapper[4948]: I0220 08:59:24.635230 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c730f22-7ac4-4c78-b897-5aed76dbbd79-utilities" (OuterVolumeSpecName: "utilities") pod "4c730f22-7ac4-4c78-b897-5aed76dbbd79" (UID: "4c730f22-7ac4-4c78-b897-5aed76dbbd79"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:59:24 crc kubenswrapper[4948]: I0220 08:59:24.642757 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c730f22-7ac4-4c78-b897-5aed76dbbd79-kube-api-access-srqhw" (OuterVolumeSpecName: "kube-api-access-srqhw") pod "4c730f22-7ac4-4c78-b897-5aed76dbbd79" (UID: "4c730f22-7ac4-4c78-b897-5aed76dbbd79"). InnerVolumeSpecName "kube-api-access-srqhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:59:24 crc kubenswrapper[4948]: I0220 08:59:24.672509 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c730f22-7ac4-4c78-b897-5aed76dbbd79-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4c730f22-7ac4-4c78-b897-5aed76dbbd79" (UID: "4c730f22-7ac4-4c78-b897-5aed76dbbd79"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:59:24 crc kubenswrapper[4948]: I0220 08:59:24.736274 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srqhw\" (UniqueName: \"kubernetes.io/projected/4c730f22-7ac4-4c78-b897-5aed76dbbd79-kube-api-access-srqhw\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:24 crc kubenswrapper[4948]: I0220 08:59:24.736607 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c730f22-7ac4-4c78-b897-5aed76dbbd79-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:24 crc kubenswrapper[4948]: I0220 08:59:24.736846 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c730f22-7ac4-4c78-b897-5aed76dbbd79-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:24 crc kubenswrapper[4948]: I0220 08:59:24.965040 4948 generic.go:334] "Generic (PLEG): container finished" podID="4c730f22-7ac4-4c78-b897-5aed76dbbd79" containerID="144c046a74ee5454148fd8c9d93123fe6564ef046b5f98222707f50a14a587be" exitCode=0 Feb 20 08:59:24 crc kubenswrapper[4948]: I0220 08:59:24.965117 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g72b5" event={"ID":"4c730f22-7ac4-4c78-b897-5aed76dbbd79","Type":"ContainerDied","Data":"144c046a74ee5454148fd8c9d93123fe6564ef046b5f98222707f50a14a587be"} Feb 20 08:59:24 crc kubenswrapper[4948]: I0220 08:59:24.965200 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g72b5" event={"ID":"4c730f22-7ac4-4c78-b897-5aed76dbbd79","Type":"ContainerDied","Data":"834d32f4c058854127072ff368549105e1f01138af7f056fc671dc63c4a176c7"} Feb 20 08:59:24 crc kubenswrapper[4948]: I0220 08:59:24.965237 4948 scope.go:117] "RemoveContainer" containerID="144c046a74ee5454148fd8c9d93123fe6564ef046b5f98222707f50a14a587be" Feb 20 08:59:24 crc kubenswrapper[4948]: I0220 08:59:24.967067 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g72b5" Feb 20 08:59:25 crc kubenswrapper[4948]: I0220 08:59:25.001059 4948 scope.go:117] "RemoveContainer" containerID="f990f9b78c70c0197487e4632df495d206960a5cc7ff502f7b84e44e7c850bf0" Feb 20 08:59:25 crc kubenswrapper[4948]: I0220 08:59:25.035547 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-g72b5"] Feb 20 08:59:25 crc kubenswrapper[4948]: I0220 08:59:25.043457 4948 scope.go:117] "RemoveContainer" containerID="f4c2df2bcbe2ce8c1b34e4d9226525be076033ebb476794a5b25b1871ed87821" Feb 20 08:59:25 crc kubenswrapper[4948]: I0220 08:59:25.057579 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-g72b5"] Feb 20 08:59:25 crc kubenswrapper[4948]: I0220 08:59:25.081367 4948 scope.go:117] "RemoveContainer" containerID="144c046a74ee5454148fd8c9d93123fe6564ef046b5f98222707f50a14a587be" Feb 20 08:59:25 crc kubenswrapper[4948]: E0220 08:59:25.081940 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"144c046a74ee5454148fd8c9d93123fe6564ef046b5f98222707f50a14a587be\": container with ID starting with 144c046a74ee5454148fd8c9d93123fe6564ef046b5f98222707f50a14a587be not found: ID does not exist" containerID="144c046a74ee5454148fd8c9d93123fe6564ef046b5f98222707f50a14a587be" Feb 20 08:59:25 crc kubenswrapper[4948]: I0220 08:59:25.082137 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"144c046a74ee5454148fd8c9d93123fe6564ef046b5f98222707f50a14a587be"} err="failed to get container status \"144c046a74ee5454148fd8c9d93123fe6564ef046b5f98222707f50a14a587be\": rpc error: code = NotFound desc = could not find container \"144c046a74ee5454148fd8c9d93123fe6564ef046b5f98222707f50a14a587be\": container with ID starting with 144c046a74ee5454148fd8c9d93123fe6564ef046b5f98222707f50a14a587be not found: ID does not exist" Feb 20 08:59:25 crc kubenswrapper[4948]: I0220 08:59:25.082292 4948 scope.go:117] "RemoveContainer" containerID="f990f9b78c70c0197487e4632df495d206960a5cc7ff502f7b84e44e7c850bf0" Feb 20 08:59:25 crc kubenswrapper[4948]: E0220 08:59:25.082782 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f990f9b78c70c0197487e4632df495d206960a5cc7ff502f7b84e44e7c850bf0\": container with ID starting with f990f9b78c70c0197487e4632df495d206960a5cc7ff502f7b84e44e7c850bf0 not found: ID does not exist" containerID="f990f9b78c70c0197487e4632df495d206960a5cc7ff502f7b84e44e7c850bf0" Feb 20 08:59:25 crc kubenswrapper[4948]: I0220 08:59:25.082930 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f990f9b78c70c0197487e4632df495d206960a5cc7ff502f7b84e44e7c850bf0"} err="failed to get container status \"f990f9b78c70c0197487e4632df495d206960a5cc7ff502f7b84e44e7c850bf0\": rpc error: code = NotFound desc = could not find container \"f990f9b78c70c0197487e4632df495d206960a5cc7ff502f7b84e44e7c850bf0\": container with ID starting with f990f9b78c70c0197487e4632df495d206960a5cc7ff502f7b84e44e7c850bf0 not found: ID does not exist" Feb 20 08:59:25 crc kubenswrapper[4948]: I0220 08:59:25.083107 4948 scope.go:117] "RemoveContainer" containerID="f4c2df2bcbe2ce8c1b34e4d9226525be076033ebb476794a5b25b1871ed87821" Feb 20 08:59:25 crc kubenswrapper[4948]: E0220 08:59:25.083611 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4c2df2bcbe2ce8c1b34e4d9226525be076033ebb476794a5b25b1871ed87821\": container with ID starting with f4c2df2bcbe2ce8c1b34e4d9226525be076033ebb476794a5b25b1871ed87821 not found: ID does not exist" containerID="f4c2df2bcbe2ce8c1b34e4d9226525be076033ebb476794a5b25b1871ed87821" Feb 20 08:59:25 crc kubenswrapper[4948]: I0220 08:59:25.083638 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4c2df2bcbe2ce8c1b34e4d9226525be076033ebb476794a5b25b1871ed87821"} err="failed to get container status \"f4c2df2bcbe2ce8c1b34e4d9226525be076033ebb476794a5b25b1871ed87821\": rpc error: code = NotFound desc = could not find container \"f4c2df2bcbe2ce8c1b34e4d9226525be076033ebb476794a5b25b1871ed87821\": container with ID starting with f4c2df2bcbe2ce8c1b34e4d9226525be076033ebb476794a5b25b1871ed87821 not found: ID does not exist" Feb 20 08:59:25 crc kubenswrapper[4948]: I0220 08:59:25.739751 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c730f22-7ac4-4c78-b897-5aed76dbbd79" path="/var/lib/kubelet/pods/4c730f22-7ac4-4c78-b897-5aed76dbbd79/volumes" Feb 20 08:59:37 crc kubenswrapper[4948]: I0220 08:59:37.875339 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-8sqsd/must-gather-xzw72"] Feb 20 08:59:37 crc kubenswrapper[4948]: E0220 08:59:37.876377 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c730f22-7ac4-4c78-b897-5aed76dbbd79" containerName="extract-utilities" Feb 20 08:59:37 crc kubenswrapper[4948]: I0220 08:59:37.876395 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c730f22-7ac4-4c78-b897-5aed76dbbd79" containerName="extract-utilities" Feb 20 08:59:37 crc kubenswrapper[4948]: E0220 08:59:37.876419 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c730f22-7ac4-4c78-b897-5aed76dbbd79" containerName="registry-server" Feb 20 08:59:37 crc kubenswrapper[4948]: I0220 08:59:37.876427 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c730f22-7ac4-4c78-b897-5aed76dbbd79" containerName="registry-server" Feb 20 08:59:37 crc kubenswrapper[4948]: E0220 08:59:37.876458 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c730f22-7ac4-4c78-b897-5aed76dbbd79" containerName="extract-content" Feb 20 08:59:37 crc kubenswrapper[4948]: I0220 08:59:37.876466 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c730f22-7ac4-4c78-b897-5aed76dbbd79" containerName="extract-content" Feb 20 08:59:37 crc kubenswrapper[4948]: I0220 08:59:37.876720 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c730f22-7ac4-4c78-b897-5aed76dbbd79" containerName="registry-server" Feb 20 08:59:37 crc kubenswrapper[4948]: I0220 08:59:37.882376 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8sqsd/must-gather-xzw72" Feb 20 08:59:37 crc kubenswrapper[4948]: I0220 08:59:37.885028 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-8sqsd"/"openshift-service-ca.crt" Feb 20 08:59:37 crc kubenswrapper[4948]: I0220 08:59:37.885310 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-8sqsd"/"kube-root-ca.crt" Feb 20 08:59:37 crc kubenswrapper[4948]: I0220 08:59:37.898738 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-8sqsd/must-gather-xzw72"] Feb 20 08:59:37 crc kubenswrapper[4948]: I0220 08:59:37.999020 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b25c36d5-5de9-4c98-acf4-495d492def68-must-gather-output\") pod \"must-gather-xzw72\" (UID: \"b25c36d5-5de9-4c98-acf4-495d492def68\") " pod="openshift-must-gather-8sqsd/must-gather-xzw72" Feb 20 08:59:37 crc kubenswrapper[4948]: I0220 08:59:37.999241 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsr8v\" (UniqueName: \"kubernetes.io/projected/b25c36d5-5de9-4c98-acf4-495d492def68-kube-api-access-vsr8v\") pod \"must-gather-xzw72\" (UID: \"b25c36d5-5de9-4c98-acf4-495d492def68\") " pod="openshift-must-gather-8sqsd/must-gather-xzw72" Feb 20 08:59:38 crc kubenswrapper[4948]: I0220 08:59:38.101526 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b25c36d5-5de9-4c98-acf4-495d492def68-must-gather-output\") pod \"must-gather-xzw72\" (UID: \"b25c36d5-5de9-4c98-acf4-495d492def68\") " pod="openshift-must-gather-8sqsd/must-gather-xzw72" Feb 20 08:59:38 crc kubenswrapper[4948]: I0220 08:59:38.101593 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsr8v\" (UniqueName: \"kubernetes.io/projected/b25c36d5-5de9-4c98-acf4-495d492def68-kube-api-access-vsr8v\") pod \"must-gather-xzw72\" (UID: \"b25c36d5-5de9-4c98-acf4-495d492def68\") " pod="openshift-must-gather-8sqsd/must-gather-xzw72" Feb 20 08:59:38 crc kubenswrapper[4948]: I0220 08:59:38.102063 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b25c36d5-5de9-4c98-acf4-495d492def68-must-gather-output\") pod \"must-gather-xzw72\" (UID: \"b25c36d5-5de9-4c98-acf4-495d492def68\") " pod="openshift-must-gather-8sqsd/must-gather-xzw72" Feb 20 08:59:38 crc kubenswrapper[4948]: I0220 08:59:38.128985 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsr8v\" (UniqueName: \"kubernetes.io/projected/b25c36d5-5de9-4c98-acf4-495d492def68-kube-api-access-vsr8v\") pod \"must-gather-xzw72\" (UID: \"b25c36d5-5de9-4c98-acf4-495d492def68\") " pod="openshift-must-gather-8sqsd/must-gather-xzw72" Feb 20 08:59:38 crc kubenswrapper[4948]: I0220 08:59:38.203168 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8sqsd/must-gather-xzw72" Feb 20 08:59:38 crc kubenswrapper[4948]: I0220 08:59:38.671399 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-8sqsd/must-gather-xzw72"] Feb 20 08:59:39 crc kubenswrapper[4948]: I0220 08:59:39.109136 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8sqsd/must-gather-xzw72" event={"ID":"b25c36d5-5de9-4c98-acf4-495d492def68","Type":"ContainerStarted","Data":"8c112f2a53188eebfa660b390f650533de90ac513fc79f81c317095abbfa670e"} Feb 20 08:59:45 crc kubenswrapper[4948]: I0220 08:59:45.180533 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8sqsd/must-gather-xzw72" event={"ID":"b25c36d5-5de9-4c98-acf4-495d492def68","Type":"ContainerStarted","Data":"ee095f93fada3e828226e407295d86f2e85f6d70e937ebcd69864d72a4cfb8b4"} Feb 20 08:59:45 crc kubenswrapper[4948]: I0220 08:59:45.181059 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8sqsd/must-gather-xzw72" event={"ID":"b25c36d5-5de9-4c98-acf4-495d492def68","Type":"ContainerStarted","Data":"4a8285325687a203e4a7cd75d1ec55e566e63219bb19f28c257e627288bb520a"} Feb 20 08:59:45 crc kubenswrapper[4948]: I0220 08:59:45.199604 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-8sqsd/must-gather-xzw72" podStartSLOduration=2.369945358 podStartE2EDuration="8.199584979s" podCreationTimestamp="2026-02-20 08:59:37 +0000 UTC" firstStartedPulling="2026-02-20 08:59:38.684707433 +0000 UTC m=+3227.659202253" lastFinishedPulling="2026-02-20 08:59:44.514347044 +0000 UTC m=+3233.488841874" observedRunningTime="2026-02-20 08:59:45.19556648 +0000 UTC m=+3234.170061300" watchObservedRunningTime="2026-02-20 08:59:45.199584979 +0000 UTC m=+3234.174079819" Feb 20 08:59:49 crc kubenswrapper[4948]: I0220 08:59:49.351644 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-8sqsd/crc-debug-2rr8w"] Feb 20 08:59:49 crc kubenswrapper[4948]: I0220 08:59:49.355613 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8sqsd/crc-debug-2rr8w" Feb 20 08:59:49 crc kubenswrapper[4948]: I0220 08:59:49.359954 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-8sqsd"/"default-dockercfg-mzzkw" Feb 20 08:59:49 crc kubenswrapper[4948]: I0220 08:59:49.522151 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bp4zr\" (UniqueName: \"kubernetes.io/projected/e1eaa8b0-3f84-4312-b7ac-4cc536093646-kube-api-access-bp4zr\") pod \"crc-debug-2rr8w\" (UID: \"e1eaa8b0-3f84-4312-b7ac-4cc536093646\") " pod="openshift-must-gather-8sqsd/crc-debug-2rr8w" Feb 20 08:59:49 crc kubenswrapper[4948]: I0220 08:59:49.522499 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e1eaa8b0-3f84-4312-b7ac-4cc536093646-host\") pod \"crc-debug-2rr8w\" (UID: \"e1eaa8b0-3f84-4312-b7ac-4cc536093646\") " pod="openshift-must-gather-8sqsd/crc-debug-2rr8w" Feb 20 08:59:49 crc kubenswrapper[4948]: I0220 08:59:49.623726 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bp4zr\" (UniqueName: \"kubernetes.io/projected/e1eaa8b0-3f84-4312-b7ac-4cc536093646-kube-api-access-bp4zr\") pod \"crc-debug-2rr8w\" (UID: \"e1eaa8b0-3f84-4312-b7ac-4cc536093646\") " pod="openshift-must-gather-8sqsd/crc-debug-2rr8w" Feb 20 08:59:49 crc kubenswrapper[4948]: I0220 08:59:49.623882 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e1eaa8b0-3f84-4312-b7ac-4cc536093646-host\") pod \"crc-debug-2rr8w\" (UID: \"e1eaa8b0-3f84-4312-b7ac-4cc536093646\") " pod="openshift-must-gather-8sqsd/crc-debug-2rr8w" Feb 20 08:59:49 crc kubenswrapper[4948]: I0220 08:59:49.623990 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e1eaa8b0-3f84-4312-b7ac-4cc536093646-host\") pod \"crc-debug-2rr8w\" (UID: \"e1eaa8b0-3f84-4312-b7ac-4cc536093646\") " pod="openshift-must-gather-8sqsd/crc-debug-2rr8w" Feb 20 08:59:49 crc kubenswrapper[4948]: I0220 08:59:49.641726 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bp4zr\" (UniqueName: \"kubernetes.io/projected/e1eaa8b0-3f84-4312-b7ac-4cc536093646-kube-api-access-bp4zr\") pod \"crc-debug-2rr8w\" (UID: \"e1eaa8b0-3f84-4312-b7ac-4cc536093646\") " pod="openshift-must-gather-8sqsd/crc-debug-2rr8w" Feb 20 08:59:49 crc kubenswrapper[4948]: I0220 08:59:49.676754 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8sqsd/crc-debug-2rr8w" Feb 20 08:59:49 crc kubenswrapper[4948]: W0220 08:59:49.711994 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1eaa8b0_3f84_4312_b7ac_4cc536093646.slice/crio-e5b803c8639b19bba9024b6977667b46b47a8a4045240c7e8378907b0fb57c9b WatchSource:0}: Error finding container e5b803c8639b19bba9024b6977667b46b47a8a4045240c7e8378907b0fb57c9b: Status 404 returned error can't find the container with id e5b803c8639b19bba9024b6977667b46b47a8a4045240c7e8378907b0fb57c9b Feb 20 08:59:50 crc kubenswrapper[4948]: I0220 08:59:50.236075 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8sqsd/crc-debug-2rr8w" event={"ID":"e1eaa8b0-3f84-4312-b7ac-4cc536093646","Type":"ContainerStarted","Data":"e5b803c8639b19bba9024b6977667b46b47a8a4045240c7e8378907b0fb57c9b"} Feb 20 09:00:00 crc kubenswrapper[4948]: I0220 09:00:00.143942 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn"] Feb 20 09:00:00 crc kubenswrapper[4948]: I0220 09:00:00.146107 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn" Feb 20 09:00:00 crc kubenswrapper[4948]: I0220 09:00:00.148665 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 20 09:00:00 crc kubenswrapper[4948]: I0220 09:00:00.149065 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 20 09:00:00 crc kubenswrapper[4948]: I0220 09:00:00.166451 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn"] Feb 20 09:00:00 crc kubenswrapper[4948]: I0220 09:00:00.220283 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ce5ff298-c730-4adc-ada7-e7a7268375e7-config-volume\") pod \"collect-profiles-29526300-g6rdn\" (UID: \"ce5ff298-c730-4adc-ada7-e7a7268375e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn" Feb 20 09:00:00 crc kubenswrapper[4948]: I0220 09:00:00.220435 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ce5ff298-c730-4adc-ada7-e7a7268375e7-secret-volume\") pod \"collect-profiles-29526300-g6rdn\" (UID: \"ce5ff298-c730-4adc-ada7-e7a7268375e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn" Feb 20 09:00:00 crc kubenswrapper[4948]: I0220 09:00:00.220610 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwqlj\" (UniqueName: \"kubernetes.io/projected/ce5ff298-c730-4adc-ada7-e7a7268375e7-kube-api-access-mwqlj\") pod \"collect-profiles-29526300-g6rdn\" (UID: \"ce5ff298-c730-4adc-ada7-e7a7268375e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn" Feb 20 09:00:00 crc kubenswrapper[4948]: I0220 09:00:00.322523 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ce5ff298-c730-4adc-ada7-e7a7268375e7-secret-volume\") pod \"collect-profiles-29526300-g6rdn\" (UID: \"ce5ff298-c730-4adc-ada7-e7a7268375e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn" Feb 20 09:00:00 crc kubenswrapper[4948]: I0220 09:00:00.322641 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwqlj\" (UniqueName: \"kubernetes.io/projected/ce5ff298-c730-4adc-ada7-e7a7268375e7-kube-api-access-mwqlj\") pod \"collect-profiles-29526300-g6rdn\" (UID: \"ce5ff298-c730-4adc-ada7-e7a7268375e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn" Feb 20 09:00:00 crc kubenswrapper[4948]: I0220 09:00:00.322744 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ce5ff298-c730-4adc-ada7-e7a7268375e7-config-volume\") pod \"collect-profiles-29526300-g6rdn\" (UID: \"ce5ff298-c730-4adc-ada7-e7a7268375e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn" Feb 20 09:00:00 crc kubenswrapper[4948]: I0220 09:00:00.323736 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ce5ff298-c730-4adc-ada7-e7a7268375e7-config-volume\") pod \"collect-profiles-29526300-g6rdn\" (UID: \"ce5ff298-c730-4adc-ada7-e7a7268375e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn" Feb 20 09:00:00 crc kubenswrapper[4948]: I0220 09:00:00.328630 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ce5ff298-c730-4adc-ada7-e7a7268375e7-secret-volume\") pod \"collect-profiles-29526300-g6rdn\" (UID: \"ce5ff298-c730-4adc-ada7-e7a7268375e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn" Feb 20 09:00:00 crc kubenswrapper[4948]: I0220 09:00:00.341520 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwqlj\" (UniqueName: \"kubernetes.io/projected/ce5ff298-c730-4adc-ada7-e7a7268375e7-kube-api-access-mwqlj\") pod \"collect-profiles-29526300-g6rdn\" (UID: \"ce5ff298-c730-4adc-ada7-e7a7268375e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn" Feb 20 09:00:00 crc kubenswrapper[4948]: I0220 09:00:00.469174 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn" Feb 20 09:00:00 crc kubenswrapper[4948]: I0220 09:00:00.944190 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn"] Feb 20 09:00:01 crc kubenswrapper[4948]: I0220 09:00:01.339185 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8sqsd/crc-debug-2rr8w" event={"ID":"e1eaa8b0-3f84-4312-b7ac-4cc536093646","Type":"ContainerStarted","Data":"d2d5efbac326ed9a4b4067346df3dcafa9289ae45429a1d4be737ea540693745"} Feb 20 09:00:01 crc kubenswrapper[4948]: I0220 09:00:01.342376 4948 generic.go:334] "Generic (PLEG): container finished" podID="ce5ff298-c730-4adc-ada7-e7a7268375e7" containerID="4e2abb2ff04ec67cba02074718775169ca5ab5ec96463e007a4d91e3fbf0ccfd" exitCode=0 Feb 20 09:00:01 crc kubenswrapper[4948]: I0220 09:00:01.342423 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn" event={"ID":"ce5ff298-c730-4adc-ada7-e7a7268375e7","Type":"ContainerDied","Data":"4e2abb2ff04ec67cba02074718775169ca5ab5ec96463e007a4d91e3fbf0ccfd"} Feb 20 09:00:01 crc kubenswrapper[4948]: I0220 09:00:01.342456 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn" event={"ID":"ce5ff298-c730-4adc-ada7-e7a7268375e7","Type":"ContainerStarted","Data":"b6a7a22636d78659185764a6d7a65810da15f8f6fec186645cce11bc9540f7fe"} Feb 20 09:00:01 crc kubenswrapper[4948]: I0220 09:00:01.362116 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-8sqsd/crc-debug-2rr8w" podStartSLOduration=1.607918326 podStartE2EDuration="12.362089958s" podCreationTimestamp="2026-02-20 08:59:49 +0000 UTC" firstStartedPulling="2026-02-20 08:59:49.715442807 +0000 UTC m=+3238.689937627" lastFinishedPulling="2026-02-20 09:00:00.469614439 +0000 UTC m=+3249.444109259" observedRunningTime="2026-02-20 09:00:01.355996587 +0000 UTC m=+3250.330491417" watchObservedRunningTime="2026-02-20 09:00:01.362089958 +0000 UTC m=+3250.336584798" Feb 20 09:00:02 crc kubenswrapper[4948]: I0220 09:00:02.761261 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn" Feb 20 09:00:02 crc kubenswrapper[4948]: I0220 09:00:02.866083 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ce5ff298-c730-4adc-ada7-e7a7268375e7-config-volume\") pod \"ce5ff298-c730-4adc-ada7-e7a7268375e7\" (UID: \"ce5ff298-c730-4adc-ada7-e7a7268375e7\") " Feb 20 09:00:02 crc kubenswrapper[4948]: I0220 09:00:02.866152 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwqlj\" (UniqueName: \"kubernetes.io/projected/ce5ff298-c730-4adc-ada7-e7a7268375e7-kube-api-access-mwqlj\") pod \"ce5ff298-c730-4adc-ada7-e7a7268375e7\" (UID: \"ce5ff298-c730-4adc-ada7-e7a7268375e7\") " Feb 20 09:00:02 crc kubenswrapper[4948]: I0220 09:00:02.866467 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ce5ff298-c730-4adc-ada7-e7a7268375e7-secret-volume\") pod \"ce5ff298-c730-4adc-ada7-e7a7268375e7\" (UID: \"ce5ff298-c730-4adc-ada7-e7a7268375e7\") " Feb 20 09:00:02 crc kubenswrapper[4948]: I0220 09:00:02.866869 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce5ff298-c730-4adc-ada7-e7a7268375e7-config-volume" (OuterVolumeSpecName: "config-volume") pod "ce5ff298-c730-4adc-ada7-e7a7268375e7" (UID: "ce5ff298-c730-4adc-ada7-e7a7268375e7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 09:00:02 crc kubenswrapper[4948]: I0220 09:00:02.868432 4948 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ce5ff298-c730-4adc-ada7-e7a7268375e7-config-volume\") on node \"crc\" DevicePath \"\"" Feb 20 09:00:02 crc kubenswrapper[4948]: I0220 09:00:02.874417 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce5ff298-c730-4adc-ada7-e7a7268375e7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ce5ff298-c730-4adc-ada7-e7a7268375e7" (UID: "ce5ff298-c730-4adc-ada7-e7a7268375e7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 09:00:02 crc kubenswrapper[4948]: I0220 09:00:02.899150 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce5ff298-c730-4adc-ada7-e7a7268375e7-kube-api-access-mwqlj" (OuterVolumeSpecName: "kube-api-access-mwqlj") pod "ce5ff298-c730-4adc-ada7-e7a7268375e7" (UID: "ce5ff298-c730-4adc-ada7-e7a7268375e7"). InnerVolumeSpecName "kube-api-access-mwqlj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:00:02 crc kubenswrapper[4948]: I0220 09:00:02.970354 4948 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ce5ff298-c730-4adc-ada7-e7a7268375e7-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 20 09:00:02 crc kubenswrapper[4948]: I0220 09:00:02.970392 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwqlj\" (UniqueName: \"kubernetes.io/projected/ce5ff298-c730-4adc-ada7-e7a7268375e7-kube-api-access-mwqlj\") on node \"crc\" DevicePath \"\"" Feb 20 09:00:03 crc kubenswrapper[4948]: I0220 09:00:03.360126 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn" event={"ID":"ce5ff298-c730-4adc-ada7-e7a7268375e7","Type":"ContainerDied","Data":"b6a7a22636d78659185764a6d7a65810da15f8f6fec186645cce11bc9540f7fe"} Feb 20 09:00:03 crc kubenswrapper[4948]: I0220 09:00:03.360175 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6a7a22636d78659185764a6d7a65810da15f8f6fec186645cce11bc9540f7fe" Feb 20 09:00:03 crc kubenswrapper[4948]: I0220 09:00:03.360148 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn" Feb 20 09:00:03 crc kubenswrapper[4948]: I0220 09:00:03.842207 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9"] Feb 20 09:00:03 crc kubenswrapper[4948]: I0220 09:00:03.849963 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9"] Feb 20 09:00:05 crc kubenswrapper[4948]: I0220 09:00:05.733382 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9db735d8-ae70-471b-9d86-3be100901c72" path="/var/lib/kubelet/pods/9db735d8-ae70-471b-9d86-3be100901c72/volumes" Feb 20 09:00:37 crc kubenswrapper[4948]: I0220 09:00:37.663226 4948 generic.go:334] "Generic (PLEG): container finished" podID="e1eaa8b0-3f84-4312-b7ac-4cc536093646" containerID="d2d5efbac326ed9a4b4067346df3dcafa9289ae45429a1d4be737ea540693745" exitCode=0 Feb 20 09:00:37 crc kubenswrapper[4948]: I0220 09:00:37.663337 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8sqsd/crc-debug-2rr8w" event={"ID":"e1eaa8b0-3f84-4312-b7ac-4cc536093646","Type":"ContainerDied","Data":"d2d5efbac326ed9a4b4067346df3dcafa9289ae45429a1d4be737ea540693745"} Feb 20 09:00:38 crc kubenswrapper[4948]: I0220 09:00:38.776126 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8sqsd/crc-debug-2rr8w" Feb 20 09:00:38 crc kubenswrapper[4948]: I0220 09:00:38.834668 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-8sqsd/crc-debug-2rr8w"] Feb 20 09:00:38 crc kubenswrapper[4948]: I0220 09:00:38.848280 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-8sqsd/crc-debug-2rr8w"] Feb 20 09:00:38 crc kubenswrapper[4948]: I0220 09:00:38.924072 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bp4zr\" (UniqueName: \"kubernetes.io/projected/e1eaa8b0-3f84-4312-b7ac-4cc536093646-kube-api-access-bp4zr\") pod \"e1eaa8b0-3f84-4312-b7ac-4cc536093646\" (UID: \"e1eaa8b0-3f84-4312-b7ac-4cc536093646\") " Feb 20 09:00:38 crc kubenswrapper[4948]: I0220 09:00:38.924123 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e1eaa8b0-3f84-4312-b7ac-4cc536093646-host\") pod \"e1eaa8b0-3f84-4312-b7ac-4cc536093646\" (UID: \"e1eaa8b0-3f84-4312-b7ac-4cc536093646\") " Feb 20 09:00:38 crc kubenswrapper[4948]: I0220 09:00:38.924545 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e1eaa8b0-3f84-4312-b7ac-4cc536093646-host" (OuterVolumeSpecName: "host") pod "e1eaa8b0-3f84-4312-b7ac-4cc536093646" (UID: "e1eaa8b0-3f84-4312-b7ac-4cc536093646"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 09:00:38 crc kubenswrapper[4948]: I0220 09:00:38.930071 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1eaa8b0-3f84-4312-b7ac-4cc536093646-kube-api-access-bp4zr" (OuterVolumeSpecName: "kube-api-access-bp4zr") pod "e1eaa8b0-3f84-4312-b7ac-4cc536093646" (UID: "e1eaa8b0-3f84-4312-b7ac-4cc536093646"). InnerVolumeSpecName "kube-api-access-bp4zr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:00:39 crc kubenswrapper[4948]: I0220 09:00:39.026407 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bp4zr\" (UniqueName: \"kubernetes.io/projected/e1eaa8b0-3f84-4312-b7ac-4cc536093646-kube-api-access-bp4zr\") on node \"crc\" DevicePath \"\"" Feb 20 09:00:39 crc kubenswrapper[4948]: I0220 09:00:39.026741 4948 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e1eaa8b0-3f84-4312-b7ac-4cc536093646-host\") on node \"crc\" DevicePath \"\"" Feb 20 09:00:39 crc kubenswrapper[4948]: I0220 09:00:39.685750 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5b803c8639b19bba9024b6977667b46b47a8a4045240c7e8378907b0fb57c9b" Feb 20 09:00:39 crc kubenswrapper[4948]: I0220 09:00:39.685808 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8sqsd/crc-debug-2rr8w" Feb 20 09:00:39 crc kubenswrapper[4948]: I0220 09:00:39.735598 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1eaa8b0-3f84-4312-b7ac-4cc536093646" path="/var/lib/kubelet/pods/e1eaa8b0-3f84-4312-b7ac-4cc536093646/volumes" Feb 20 09:00:40 crc kubenswrapper[4948]: I0220 09:00:40.020601 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-8sqsd/crc-debug-fnl8n"] Feb 20 09:00:40 crc kubenswrapper[4948]: E0220 09:00:40.021436 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1eaa8b0-3f84-4312-b7ac-4cc536093646" containerName="container-00" Feb 20 09:00:40 crc kubenswrapper[4948]: I0220 09:00:40.021453 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1eaa8b0-3f84-4312-b7ac-4cc536093646" containerName="container-00" Feb 20 09:00:40 crc kubenswrapper[4948]: E0220 09:00:40.021477 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce5ff298-c730-4adc-ada7-e7a7268375e7" containerName="collect-profiles" Feb 20 09:00:40 crc kubenswrapper[4948]: I0220 09:00:40.021485 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce5ff298-c730-4adc-ada7-e7a7268375e7" containerName="collect-profiles" Feb 20 09:00:40 crc kubenswrapper[4948]: I0220 09:00:40.021715 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce5ff298-c730-4adc-ada7-e7a7268375e7" containerName="collect-profiles" Feb 20 09:00:40 crc kubenswrapper[4948]: I0220 09:00:40.021735 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1eaa8b0-3f84-4312-b7ac-4cc536093646" containerName="container-00" Feb 20 09:00:40 crc kubenswrapper[4948]: I0220 09:00:40.022450 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8sqsd/crc-debug-fnl8n" Feb 20 09:00:40 crc kubenswrapper[4948]: I0220 09:00:40.024921 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-8sqsd"/"default-dockercfg-mzzkw" Feb 20 09:00:40 crc kubenswrapper[4948]: I0220 09:00:40.147605 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ngvx\" (UniqueName: \"kubernetes.io/projected/2116283d-53f0-45f3-9d0a-43f1af3d31d3-kube-api-access-8ngvx\") pod \"crc-debug-fnl8n\" (UID: \"2116283d-53f0-45f3-9d0a-43f1af3d31d3\") " pod="openshift-must-gather-8sqsd/crc-debug-fnl8n" Feb 20 09:00:40 crc kubenswrapper[4948]: I0220 09:00:40.147666 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2116283d-53f0-45f3-9d0a-43f1af3d31d3-host\") pod \"crc-debug-fnl8n\" (UID: \"2116283d-53f0-45f3-9d0a-43f1af3d31d3\") " pod="openshift-must-gather-8sqsd/crc-debug-fnl8n" Feb 20 09:00:40 crc kubenswrapper[4948]: I0220 09:00:40.249648 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ngvx\" (UniqueName: \"kubernetes.io/projected/2116283d-53f0-45f3-9d0a-43f1af3d31d3-kube-api-access-8ngvx\") pod \"crc-debug-fnl8n\" (UID: \"2116283d-53f0-45f3-9d0a-43f1af3d31d3\") " pod="openshift-must-gather-8sqsd/crc-debug-fnl8n" Feb 20 09:00:40 crc kubenswrapper[4948]: I0220 09:00:40.249702 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2116283d-53f0-45f3-9d0a-43f1af3d31d3-host\") pod \"crc-debug-fnl8n\" (UID: \"2116283d-53f0-45f3-9d0a-43f1af3d31d3\") " pod="openshift-must-gather-8sqsd/crc-debug-fnl8n" Feb 20 09:00:40 crc kubenswrapper[4948]: I0220 09:00:40.249855 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2116283d-53f0-45f3-9d0a-43f1af3d31d3-host\") pod \"crc-debug-fnl8n\" (UID: \"2116283d-53f0-45f3-9d0a-43f1af3d31d3\") " pod="openshift-must-gather-8sqsd/crc-debug-fnl8n" Feb 20 09:00:40 crc kubenswrapper[4948]: I0220 09:00:40.274358 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ngvx\" (UniqueName: \"kubernetes.io/projected/2116283d-53f0-45f3-9d0a-43f1af3d31d3-kube-api-access-8ngvx\") pod \"crc-debug-fnl8n\" (UID: \"2116283d-53f0-45f3-9d0a-43f1af3d31d3\") " pod="openshift-must-gather-8sqsd/crc-debug-fnl8n" Feb 20 09:00:40 crc kubenswrapper[4948]: I0220 09:00:40.337577 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8sqsd/crc-debug-fnl8n" Feb 20 09:00:40 crc kubenswrapper[4948]: I0220 09:00:40.695855 4948 generic.go:334] "Generic (PLEG): container finished" podID="2116283d-53f0-45f3-9d0a-43f1af3d31d3" containerID="c7e69ff578f0981df401ec9cf440f5ada90becff9fd036ae8616d13d44ef4205" exitCode=0 Feb 20 09:00:40 crc kubenswrapper[4948]: I0220 09:00:40.695931 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8sqsd/crc-debug-fnl8n" event={"ID":"2116283d-53f0-45f3-9d0a-43f1af3d31d3","Type":"ContainerDied","Data":"c7e69ff578f0981df401ec9cf440f5ada90becff9fd036ae8616d13d44ef4205"} Feb 20 09:00:40 crc kubenswrapper[4948]: I0220 09:00:40.696233 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8sqsd/crc-debug-fnl8n" event={"ID":"2116283d-53f0-45f3-9d0a-43f1af3d31d3","Type":"ContainerStarted","Data":"81333fd279dddbfd5f99a192bc9a5787fce10efd0531cd1b913b8e7cb477ce51"} Feb 20 09:00:41 crc kubenswrapper[4948]: I0220 09:00:41.108301 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-8sqsd/crc-debug-fnl8n"] Feb 20 09:00:41 crc kubenswrapper[4948]: I0220 09:00:41.118115 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-8sqsd/crc-debug-fnl8n"] Feb 20 09:00:41 crc kubenswrapper[4948]: I0220 09:00:41.815415 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8sqsd/crc-debug-fnl8n" Feb 20 09:00:41 crc kubenswrapper[4948]: I0220 09:00:41.977848 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2116283d-53f0-45f3-9d0a-43f1af3d31d3-host\") pod \"2116283d-53f0-45f3-9d0a-43f1af3d31d3\" (UID: \"2116283d-53f0-45f3-9d0a-43f1af3d31d3\") " Feb 20 09:00:41 crc kubenswrapper[4948]: I0220 09:00:41.978043 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ngvx\" (UniqueName: \"kubernetes.io/projected/2116283d-53f0-45f3-9d0a-43f1af3d31d3-kube-api-access-8ngvx\") pod \"2116283d-53f0-45f3-9d0a-43f1af3d31d3\" (UID: \"2116283d-53f0-45f3-9d0a-43f1af3d31d3\") " Feb 20 09:00:41 crc kubenswrapper[4948]: I0220 09:00:41.977955 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2116283d-53f0-45f3-9d0a-43f1af3d31d3-host" (OuterVolumeSpecName: "host") pod "2116283d-53f0-45f3-9d0a-43f1af3d31d3" (UID: "2116283d-53f0-45f3-9d0a-43f1af3d31d3"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 09:00:41 crc kubenswrapper[4948]: I0220 09:00:41.978504 4948 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2116283d-53f0-45f3-9d0a-43f1af3d31d3-host\") on node \"crc\" DevicePath \"\"" Feb 20 09:00:41 crc kubenswrapper[4948]: I0220 09:00:41.982807 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2116283d-53f0-45f3-9d0a-43f1af3d31d3-kube-api-access-8ngvx" (OuterVolumeSpecName: "kube-api-access-8ngvx") pod "2116283d-53f0-45f3-9d0a-43f1af3d31d3" (UID: "2116283d-53f0-45f3-9d0a-43f1af3d31d3"). InnerVolumeSpecName "kube-api-access-8ngvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:00:42 crc kubenswrapper[4948]: I0220 09:00:42.080434 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ngvx\" (UniqueName: \"kubernetes.io/projected/2116283d-53f0-45f3-9d0a-43f1af3d31d3-kube-api-access-8ngvx\") on node \"crc\" DevicePath \"\"" Feb 20 09:00:42 crc kubenswrapper[4948]: I0220 09:00:42.291855 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-8sqsd/crc-debug-dbvll"] Feb 20 09:00:42 crc kubenswrapper[4948]: E0220 09:00:42.293352 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2116283d-53f0-45f3-9d0a-43f1af3d31d3" containerName="container-00" Feb 20 09:00:42 crc kubenswrapper[4948]: I0220 09:00:42.293484 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="2116283d-53f0-45f3-9d0a-43f1af3d31d3" containerName="container-00" Feb 20 09:00:42 crc kubenswrapper[4948]: I0220 09:00:42.293819 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="2116283d-53f0-45f3-9d0a-43f1af3d31d3" containerName="container-00" Feb 20 09:00:42 crc kubenswrapper[4948]: I0220 09:00:42.294651 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8sqsd/crc-debug-dbvll" Feb 20 09:00:42 crc kubenswrapper[4948]: I0220 09:00:42.387036 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7f9bq\" (UniqueName: \"kubernetes.io/projected/e9fba2ec-0e22-4b23-aaa6-0406521ff59e-kube-api-access-7f9bq\") pod \"crc-debug-dbvll\" (UID: \"e9fba2ec-0e22-4b23-aaa6-0406521ff59e\") " pod="openshift-must-gather-8sqsd/crc-debug-dbvll" Feb 20 09:00:42 crc kubenswrapper[4948]: I0220 09:00:42.387354 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e9fba2ec-0e22-4b23-aaa6-0406521ff59e-host\") pod \"crc-debug-dbvll\" (UID: \"e9fba2ec-0e22-4b23-aaa6-0406521ff59e\") " pod="openshift-must-gather-8sqsd/crc-debug-dbvll" Feb 20 09:00:42 crc kubenswrapper[4948]: I0220 09:00:42.489555 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7f9bq\" (UniqueName: \"kubernetes.io/projected/e9fba2ec-0e22-4b23-aaa6-0406521ff59e-kube-api-access-7f9bq\") pod \"crc-debug-dbvll\" (UID: \"e9fba2ec-0e22-4b23-aaa6-0406521ff59e\") " pod="openshift-must-gather-8sqsd/crc-debug-dbvll" Feb 20 09:00:42 crc kubenswrapper[4948]: I0220 09:00:42.489647 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e9fba2ec-0e22-4b23-aaa6-0406521ff59e-host\") pod \"crc-debug-dbvll\" (UID: \"e9fba2ec-0e22-4b23-aaa6-0406521ff59e\") " pod="openshift-must-gather-8sqsd/crc-debug-dbvll" Feb 20 09:00:42 crc kubenswrapper[4948]: I0220 09:00:42.489831 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e9fba2ec-0e22-4b23-aaa6-0406521ff59e-host\") pod \"crc-debug-dbvll\" (UID: \"e9fba2ec-0e22-4b23-aaa6-0406521ff59e\") " pod="openshift-must-gather-8sqsd/crc-debug-dbvll" Feb 20 09:00:42 crc kubenswrapper[4948]: I0220 09:00:42.505248 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7f9bq\" (UniqueName: \"kubernetes.io/projected/e9fba2ec-0e22-4b23-aaa6-0406521ff59e-kube-api-access-7f9bq\") pod \"crc-debug-dbvll\" (UID: \"e9fba2ec-0e22-4b23-aaa6-0406521ff59e\") " pod="openshift-must-gather-8sqsd/crc-debug-dbvll" Feb 20 09:00:42 crc kubenswrapper[4948]: I0220 09:00:42.609661 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8sqsd/crc-debug-dbvll" Feb 20 09:00:42 crc kubenswrapper[4948]: W0220 09:00:42.631820 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9fba2ec_0e22_4b23_aaa6_0406521ff59e.slice/crio-ffaf11f2cf3db38a687d2ffbc7c361de400b4facf9ae2cf5417b1b215c46e524 WatchSource:0}: Error finding container ffaf11f2cf3db38a687d2ffbc7c361de400b4facf9ae2cf5417b1b215c46e524: Status 404 returned error can't find the container with id ffaf11f2cf3db38a687d2ffbc7c361de400b4facf9ae2cf5417b1b215c46e524 Feb 20 09:00:42 crc kubenswrapper[4948]: I0220 09:00:42.713106 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8sqsd/crc-debug-dbvll" event={"ID":"e9fba2ec-0e22-4b23-aaa6-0406521ff59e","Type":"ContainerStarted","Data":"ffaf11f2cf3db38a687d2ffbc7c361de400b4facf9ae2cf5417b1b215c46e524"} Feb 20 09:00:42 crc kubenswrapper[4948]: I0220 09:00:42.715035 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8sqsd/crc-debug-fnl8n" Feb 20 09:00:42 crc kubenswrapper[4948]: I0220 09:00:42.715031 4948 scope.go:117] "RemoveContainer" containerID="c7e69ff578f0981df401ec9cf440f5ada90becff9fd036ae8616d13d44ef4205" Feb 20 09:00:43 crc kubenswrapper[4948]: I0220 09:00:43.725409 4948 generic.go:334] "Generic (PLEG): container finished" podID="e9fba2ec-0e22-4b23-aaa6-0406521ff59e" containerID="a21d252ab88ad9dc35d3fb9b8bd4786641c231e5c515a0b9849eb5e01174f261" exitCode=0 Feb 20 09:00:43 crc kubenswrapper[4948]: I0220 09:00:43.738927 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2116283d-53f0-45f3-9d0a-43f1af3d31d3" path="/var/lib/kubelet/pods/2116283d-53f0-45f3-9d0a-43f1af3d31d3/volumes" Feb 20 09:00:43 crc kubenswrapper[4948]: I0220 09:00:43.739774 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8sqsd/crc-debug-dbvll" event={"ID":"e9fba2ec-0e22-4b23-aaa6-0406521ff59e","Type":"ContainerDied","Data":"a21d252ab88ad9dc35d3fb9b8bd4786641c231e5c515a0b9849eb5e01174f261"} Feb 20 09:00:43 crc kubenswrapper[4948]: I0220 09:00:43.775814 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-8sqsd/crc-debug-dbvll"] Feb 20 09:00:43 crc kubenswrapper[4948]: I0220 09:00:43.788650 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-8sqsd/crc-debug-dbvll"] Feb 20 09:00:44 crc kubenswrapper[4948]: I0220 09:00:44.854095 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8sqsd/crc-debug-dbvll" Feb 20 09:00:44 crc kubenswrapper[4948]: I0220 09:00:44.940875 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e9fba2ec-0e22-4b23-aaa6-0406521ff59e-host\") pod \"e9fba2ec-0e22-4b23-aaa6-0406521ff59e\" (UID: \"e9fba2ec-0e22-4b23-aaa6-0406521ff59e\") " Feb 20 09:00:44 crc kubenswrapper[4948]: I0220 09:00:44.940964 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7f9bq\" (UniqueName: \"kubernetes.io/projected/e9fba2ec-0e22-4b23-aaa6-0406521ff59e-kube-api-access-7f9bq\") pod \"e9fba2ec-0e22-4b23-aaa6-0406521ff59e\" (UID: \"e9fba2ec-0e22-4b23-aaa6-0406521ff59e\") " Feb 20 09:00:44 crc kubenswrapper[4948]: I0220 09:00:44.941030 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e9fba2ec-0e22-4b23-aaa6-0406521ff59e-host" (OuterVolumeSpecName: "host") pod "e9fba2ec-0e22-4b23-aaa6-0406521ff59e" (UID: "e9fba2ec-0e22-4b23-aaa6-0406521ff59e"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 09:00:44 crc kubenswrapper[4948]: I0220 09:00:44.941579 4948 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e9fba2ec-0e22-4b23-aaa6-0406521ff59e-host\") on node \"crc\" DevicePath \"\"" Feb 20 09:00:44 crc kubenswrapper[4948]: I0220 09:00:44.955224 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9fba2ec-0e22-4b23-aaa6-0406521ff59e-kube-api-access-7f9bq" (OuterVolumeSpecName: "kube-api-access-7f9bq") pod "e9fba2ec-0e22-4b23-aaa6-0406521ff59e" (UID: "e9fba2ec-0e22-4b23-aaa6-0406521ff59e"). InnerVolumeSpecName "kube-api-access-7f9bq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:00:45 crc kubenswrapper[4948]: I0220 09:00:45.043282 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7f9bq\" (UniqueName: \"kubernetes.io/projected/e9fba2ec-0e22-4b23-aaa6-0406521ff59e-kube-api-access-7f9bq\") on node \"crc\" DevicePath \"\"" Feb 20 09:00:45 crc kubenswrapper[4948]: I0220 09:00:45.732786 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9fba2ec-0e22-4b23-aaa6-0406521ff59e" path="/var/lib/kubelet/pods/e9fba2ec-0e22-4b23-aaa6-0406521ff59e/volumes" Feb 20 09:00:45 crc kubenswrapper[4948]: I0220 09:00:45.747583 4948 scope.go:117] "RemoveContainer" containerID="a21d252ab88ad9dc35d3fb9b8bd4786641c231e5c515a0b9849eb5e01174f261" Feb 20 09:00:45 crc kubenswrapper[4948]: I0220 09:00:45.747608 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8sqsd/crc-debug-dbvll" Feb 20 09:00:59 crc kubenswrapper[4948]: I0220 09:00:59.172307 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-686d666564-8tnjb_51f6301c-57be-4a07-91df-b1a9f90e3f40/barbican-api/0.log" Feb 20 09:00:59 crc kubenswrapper[4948]: I0220 09:00:59.325096 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-686d666564-8tnjb_51f6301c-57be-4a07-91df-b1a9f90e3f40/barbican-api-log/0.log" Feb 20 09:00:59 crc kubenswrapper[4948]: I0220 09:00:59.398017 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6ffd468-xgk27_292d8e1e-de3b-4259-afef-c19e2c21187c/barbican-keystone-listener/0.log" Feb 20 09:00:59 crc kubenswrapper[4948]: I0220 09:00:59.438356 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6ffd468-xgk27_292d8e1e-de3b-4259-afef-c19e2c21187c/barbican-keystone-listener-log/0.log" Feb 20 09:00:59 crc kubenswrapper[4948]: I0220 09:00:59.571773 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-76d486c65f-w9lc6_4b356058-d957-4fa5-a106-3fe5052c8539/barbican-worker/0.log" Feb 20 09:00:59 crc kubenswrapper[4948]: I0220 09:00:59.585502 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-76d486c65f-w9lc6_4b356058-d957-4fa5-a106-3fe5052c8539/barbican-worker-log/0.log" Feb 20 09:00:59 crc kubenswrapper[4948]: I0220 09:00:59.742327 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-4265g_e9bed94f-dc85-433b-b2db-8da400959f54/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:00:59 crc kubenswrapper[4948]: I0220 09:00:59.792003 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_58f563bb-9716-4abd-b187-3f01036d4f31/ceilometer-central-agent/0.log" Feb 20 09:00:59 crc kubenswrapper[4948]: I0220 09:00:59.883258 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_58f563bb-9716-4abd-b187-3f01036d4f31/ceilometer-notification-agent/0.log" Feb 20 09:00:59 crc kubenswrapper[4948]: I0220 09:00:59.906626 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_58f563bb-9716-4abd-b187-3f01036d4f31/proxy-httpd/0.log" Feb 20 09:00:59 crc kubenswrapper[4948]: I0220 09:00:59.936912 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_58f563bb-9716-4abd-b187-3f01036d4f31/sg-core/0.log" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.069874 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_e5967cab-cc29-4c87-b873-0f5d20203d4f/cinder-api/0.log" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.105335 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_e5967cab-cc29-4c87-b873-0f5d20203d4f/cinder-api-log/0.log" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.146472 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29526301-gklw9"] Feb 20 09:01:00 crc kubenswrapper[4948]: E0220 09:01:00.147088 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9fba2ec-0e22-4b23-aaa6-0406521ff59e" containerName="container-00" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.147121 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9fba2ec-0e22-4b23-aaa6-0406521ff59e" containerName="container-00" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.147423 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9fba2ec-0e22-4b23-aaa6-0406521ff59e" containerName="container-00" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.148218 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29526301-gklw9" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.186273 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29526301-gklw9"] Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.259634 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-combined-ca-bundle\") pod \"keystone-cron-29526301-gklw9\" (UID: \"7990f62f-cd0f-4be2-b166-c00621d877e1\") " pod="openstack/keystone-cron-29526301-gklw9" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.259766 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxrlv\" (UniqueName: \"kubernetes.io/projected/7990f62f-cd0f-4be2-b166-c00621d877e1-kube-api-access-pxrlv\") pod \"keystone-cron-29526301-gklw9\" (UID: \"7990f62f-cd0f-4be2-b166-c00621d877e1\") " pod="openstack/keystone-cron-29526301-gklw9" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.259864 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-fernet-keys\") pod \"keystone-cron-29526301-gklw9\" (UID: \"7990f62f-cd0f-4be2-b166-c00621d877e1\") " pod="openstack/keystone-cron-29526301-gklw9" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.259913 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-config-data\") pod \"keystone-cron-29526301-gklw9\" (UID: \"7990f62f-cd0f-4be2-b166-c00621d877e1\") " pod="openstack/keystone-cron-29526301-gklw9" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.344331 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2f8a0f7b-ee05-4e93-a591-7486fecf8cd0/probe/0.log" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.346092 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2f8a0f7b-ee05-4e93-a591-7486fecf8cd0/cinder-scheduler/0.log" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.361046 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxrlv\" (UniqueName: \"kubernetes.io/projected/7990f62f-cd0f-4be2-b166-c00621d877e1-kube-api-access-pxrlv\") pod \"keystone-cron-29526301-gklw9\" (UID: \"7990f62f-cd0f-4be2-b166-c00621d877e1\") " pod="openstack/keystone-cron-29526301-gklw9" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.361115 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-fernet-keys\") pod \"keystone-cron-29526301-gklw9\" (UID: \"7990f62f-cd0f-4be2-b166-c00621d877e1\") " pod="openstack/keystone-cron-29526301-gklw9" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.361150 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-config-data\") pod \"keystone-cron-29526301-gklw9\" (UID: \"7990f62f-cd0f-4be2-b166-c00621d877e1\") " pod="openstack/keystone-cron-29526301-gklw9" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.361225 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-combined-ca-bundle\") pod \"keystone-cron-29526301-gklw9\" (UID: \"7990f62f-cd0f-4be2-b166-c00621d877e1\") " pod="openstack/keystone-cron-29526301-gklw9" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.367023 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-combined-ca-bundle\") pod \"keystone-cron-29526301-gklw9\" (UID: \"7990f62f-cd0f-4be2-b166-c00621d877e1\") " pod="openstack/keystone-cron-29526301-gklw9" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.369401 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-config-data\") pod \"keystone-cron-29526301-gklw9\" (UID: \"7990f62f-cd0f-4be2-b166-c00621d877e1\") " pod="openstack/keystone-cron-29526301-gklw9" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.381502 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-fernet-keys\") pod \"keystone-cron-29526301-gklw9\" (UID: \"7990f62f-cd0f-4be2-b166-c00621d877e1\") " pod="openstack/keystone-cron-29526301-gklw9" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.381875 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxrlv\" (UniqueName: \"kubernetes.io/projected/7990f62f-cd0f-4be2-b166-c00621d877e1-kube-api-access-pxrlv\") pod \"keystone-cron-29526301-gklw9\" (UID: \"7990f62f-cd0f-4be2-b166-c00621d877e1\") " pod="openstack/keystone-cron-29526301-gklw9" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.390939 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx_bf3501bf-8f1e-4529-91a8-6be83eda4158/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.513359 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29526301-gklw9" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.622314 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-fbjpk_b05f92c6-60c9-41be-bf6c-6288cee04659/init/0.log" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.642433 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6_d38195a4-0c7b-427e-b5e2-923b24c10674/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.818224 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-fbjpk_b05f92c6-60c9-41be-bf6c-6288cee04659/init/0.log" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.896363 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-mhkll_5864a68d-650e-4bcf-b705-619c0f27445b/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.899506 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-fbjpk_b05f92c6-60c9-41be-bf6c-6288cee04659/dnsmasq-dns/0.log" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.961305 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29526301-gklw9"] Feb 20 09:01:01 crc kubenswrapper[4948]: I0220 09:01:01.086848 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a/glance-log/0.log" Feb 20 09:01:01 crc kubenswrapper[4948]: I0220 09:01:01.116530 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a/glance-httpd/0.log" Feb 20 09:01:01 crc kubenswrapper[4948]: I0220 09:01:01.264467 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_4fcb965e-4d04-4863-a966-39a83f458fa6/glance-httpd/0.log" Feb 20 09:01:01 crc kubenswrapper[4948]: I0220 09:01:01.281574 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_4fcb965e-4d04-4863-a966-39a83f458fa6/glance-log/0.log" Feb 20 09:01:01 crc kubenswrapper[4948]: I0220 09:01:01.427722 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-685c4db87b-rw57h_7b9a072a-a040-414c-96c7-cf1454148745/horizon/0.log" Feb 20 09:01:01 crc kubenswrapper[4948]: I0220 09:01:01.592489 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw_c517042a-5354-4a8d-b7ff-c9aafa263b6c/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:01:01 crc kubenswrapper[4948]: I0220 09:01:01.694513 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-685c4db87b-rw57h_7b9a072a-a040-414c-96c7-cf1454148745/horizon-log/0.log" Feb 20 09:01:01 crc kubenswrapper[4948]: I0220 09:01:01.765884 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-4l55v_f42ddd0c-a2ec-414a-812d-f87213a3226d/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:01:01 crc kubenswrapper[4948]: I0220 09:01:01.883543 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29526301-gklw9" event={"ID":"7990f62f-cd0f-4be2-b166-c00621d877e1","Type":"ContainerStarted","Data":"62da2ba3fadb7d2ed292d6ef544e7a030f473562d41eff9af8ce7a20a18cc887"} Feb 20 09:01:01 crc kubenswrapper[4948]: I0220 09:01:01.883593 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29526301-gklw9" event={"ID":"7990f62f-cd0f-4be2-b166-c00621d877e1","Type":"ContainerStarted","Data":"354be5b8ab4377e115ef71435a00840f8779e3587ce24ad7e77f1b79ea13775a"} Feb 20 09:01:01 crc kubenswrapper[4948]: I0220 09:01:01.905262 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29526301-gklw9" podStartSLOduration=1.905245212 podStartE2EDuration="1.905245212s" podCreationTimestamp="2026-02-20 09:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 09:01:01.900883974 +0000 UTC m=+3310.875378794" watchObservedRunningTime="2026-02-20 09:01:01.905245212 +0000 UTC m=+3310.879740032" Feb 20 09:01:01 crc kubenswrapper[4948]: I0220 09:01:01.996915 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_24bf0fb4-81db-4955-8b5f-92c0f6b0856b/kube-state-metrics/0.log" Feb 20 09:01:02 crc kubenswrapper[4948]: I0220 09:01:02.041128 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6d4b8df464-wspg2_6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc/keystone-api/0.log" Feb 20 09:01:02 crc kubenswrapper[4948]: I0220 09:01:02.290214 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs_6e619eac-dfdc-4c8c-aa51-fc08ba58b49a/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:01:02 crc kubenswrapper[4948]: I0220 09:01:02.660619 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-76b5684765-h5624_fb9d5a3e-b07f-4229-9b5b-d0db6750cf59/neutron-api/0.log" Feb 20 09:01:02 crc kubenswrapper[4948]: I0220 09:01:02.786752 4948 scope.go:117] "RemoveContainer" containerID="984bcec5b88d4682af52feec68d576f83622efc6d9578f02504c68786eff5cfd" Feb 20 09:01:02 crc kubenswrapper[4948]: I0220 09:01:02.803357 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-76b5684765-h5624_fb9d5a3e-b07f-4229-9b5b-d0db6750cf59/neutron-httpd/0.log" Feb 20 09:01:02 crc kubenswrapper[4948]: I0220 09:01:02.942448 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr_fd280342-4276-4eee-a763-876e31be28c5/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:01:03 crc kubenswrapper[4948]: I0220 09:01:03.504550 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_873a6553-9637-4f2a-a743-f33cde7fc883/nova-cell0-conductor-conductor/0.log" Feb 20 09:01:03 crc kubenswrapper[4948]: I0220 09:01:03.506508 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_354f9b2f-9eaa-4e47-b214-0fd895f90c4c/nova-api-log/0.log" Feb 20 09:01:03 crc kubenswrapper[4948]: I0220 09:01:03.664898 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_354f9b2f-9eaa-4e47-b214-0fd895f90c4c/nova-api-api/0.log" Feb 20 09:01:03 crc kubenswrapper[4948]: I0220 09:01:03.834059 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_e801ada9-0247-4b29-b262-04637e1f8452/nova-cell1-conductor-conductor/0.log" Feb 20 09:01:03 crc kubenswrapper[4948]: I0220 09:01:03.853702 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_9c00f24b-fedb-4747-91c4-3d0b551a7288/nova-cell1-novncproxy-novncproxy/0.log" Feb 20 09:01:03 crc kubenswrapper[4948]: I0220 09:01:03.907027 4948 generic.go:334] "Generic (PLEG): container finished" podID="7990f62f-cd0f-4be2-b166-c00621d877e1" containerID="62da2ba3fadb7d2ed292d6ef544e7a030f473562d41eff9af8ce7a20a18cc887" exitCode=0 Feb 20 09:01:03 crc kubenswrapper[4948]: I0220 09:01:03.907067 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29526301-gklw9" event={"ID":"7990f62f-cd0f-4be2-b166-c00621d877e1","Type":"ContainerDied","Data":"62da2ba3fadb7d2ed292d6ef544e7a030f473562d41eff9af8ce7a20a18cc887"} Feb 20 09:01:04 crc kubenswrapper[4948]: I0220 09:01:04.076170 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-p22xr_6005ded0-9e21-4908-bbf0-33c710ba4341/nova-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:01:04 crc kubenswrapper[4948]: I0220 09:01:04.148891 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_59965363-9971-4130-bb5a-5fcf5a44e2d5/nova-metadata-log/0.log" Feb 20 09:01:04 crc kubenswrapper[4948]: I0220 09:01:04.470643 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_f114bc41-01a5-4955-97eb-7fcf139cc5a9/nova-scheduler-scheduler/0.log" Feb 20 09:01:04 crc kubenswrapper[4948]: I0220 09:01:04.486124 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_d02114ae-21f2-4cc5-9bca-9505c25eaf5f/mysql-bootstrap/0.log" Feb 20 09:01:04 crc kubenswrapper[4948]: I0220 09:01:04.650660 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_d02114ae-21f2-4cc5-9bca-9505c25eaf5f/mysql-bootstrap/0.log" Feb 20 09:01:04 crc kubenswrapper[4948]: I0220 09:01:04.670220 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_d02114ae-21f2-4cc5-9bca-9505c25eaf5f/galera/0.log" Feb 20 09:01:04 crc kubenswrapper[4948]: I0220 09:01:04.868310 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_6e59b63d-0c56-488c-87cb-348af87058c6/mysql-bootstrap/0.log" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.141141 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_6e59b63d-0c56-488c-87cb-348af87058c6/mysql-bootstrap/0.log" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.148557 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_6e59b63d-0c56-488c-87cb-348af87058c6/galera/0.log" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.237881 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29526301-gklw9" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.244427 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_59965363-9971-4130-bb5a-5fcf5a44e2d5/nova-metadata-metadata/0.log" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.315043 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_b2a76f68-3b4b-4993-aa4f-3b31b4ee1008/openstackclient/0.log" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.357187 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxrlv\" (UniqueName: \"kubernetes.io/projected/7990f62f-cd0f-4be2-b166-c00621d877e1-kube-api-access-pxrlv\") pod \"7990f62f-cd0f-4be2-b166-c00621d877e1\" (UID: \"7990f62f-cd0f-4be2-b166-c00621d877e1\") " Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.357248 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-fernet-keys\") pod \"7990f62f-cd0f-4be2-b166-c00621d877e1\" (UID: \"7990f62f-cd0f-4be2-b166-c00621d877e1\") " Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.357331 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-combined-ca-bundle\") pod \"7990f62f-cd0f-4be2-b166-c00621d877e1\" (UID: \"7990f62f-cd0f-4be2-b166-c00621d877e1\") " Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.357365 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-config-data\") pod \"7990f62f-cd0f-4be2-b166-c00621d877e1\" (UID: \"7990f62f-cd0f-4be2-b166-c00621d877e1\") " Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.375389 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7990f62f-cd0f-4be2-b166-c00621d877e1-kube-api-access-pxrlv" (OuterVolumeSpecName: "kube-api-access-pxrlv") pod "7990f62f-cd0f-4be2-b166-c00621d877e1" (UID: "7990f62f-cd0f-4be2-b166-c00621d877e1"). InnerVolumeSpecName "kube-api-access-pxrlv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.375489 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "7990f62f-cd0f-4be2-b166-c00621d877e1" (UID: "7990f62f-cd0f-4be2-b166-c00621d877e1"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.411073 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-config-data" (OuterVolumeSpecName: "config-data") pod "7990f62f-cd0f-4be2-b166-c00621d877e1" (UID: "7990f62f-cd0f-4be2-b166-c00621d877e1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.446180 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-gcfn8_1510e8b2-af6f-4641-a2d2-361830382250/openstack-network-exporter/0.log" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.460786 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.460811 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxrlv\" (UniqueName: \"kubernetes.io/projected/7990f62f-cd0f-4be2-b166-c00621d877e1-kube-api-access-pxrlv\") on node \"crc\" DevicePath \"\"" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.460821 4948 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.478298 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7990f62f-cd0f-4be2-b166-c00621d877e1" (UID: "7990f62f-cd0f-4be2-b166-c00621d877e1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.563092 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.756277 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-q4bv4_c82772b6-9737-45ee-998a-e897086c03b5/ovsdb-server-init/0.log" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.766266 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-mkpng_b80cb988-de2a-4e65-9161-a0af0561c754/ovn-controller/0.log" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.922855 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-q4bv4_c82772b6-9737-45ee-998a-e897086c03b5/ovsdb-server-init/0.log" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.926376 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29526301-gklw9" event={"ID":"7990f62f-cd0f-4be2-b166-c00621d877e1","Type":"ContainerDied","Data":"354be5b8ab4377e115ef71435a00840f8779e3587ce24ad7e77f1b79ea13775a"} Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.926414 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="354be5b8ab4377e115ef71435a00840f8779e3587ce24ad7e77f1b79ea13775a" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.926467 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29526301-gklw9" Feb 20 09:01:06 crc kubenswrapper[4948]: I0220 09:01:06.025946 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-q4bv4_c82772b6-9737-45ee-998a-e897086c03b5/ovs-vswitchd/0.log" Feb 20 09:01:06 crc kubenswrapper[4948]: I0220 09:01:06.072718 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-q4bv4_c82772b6-9737-45ee-998a-e897086c03b5/ovsdb-server/0.log" Feb 20 09:01:06 crc kubenswrapper[4948]: I0220 09:01:06.152071 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-4b2vp_b19ca7cb-01d2-4965-90f6-5d50211870cc/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:01:06 crc kubenswrapper[4948]: I0220 09:01:06.234052 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_eeb3782c-0a88-4d17-b481-e2aef7ea386a/openstack-network-exporter/0.log" Feb 20 09:01:06 crc kubenswrapper[4948]: I0220 09:01:06.280569 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_eeb3782c-0a88-4d17-b481-e2aef7ea386a/ovn-northd/0.log" Feb 20 09:01:06 crc kubenswrapper[4948]: I0220 09:01:06.424119 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_d681442e-6bdb-46e4-9b49-6466e5e036a0/openstack-network-exporter/0.log" Feb 20 09:01:06 crc kubenswrapper[4948]: I0220 09:01:06.449054 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_d681442e-6bdb-46e4-9b49-6466e5e036a0/ovsdbserver-nb/0.log" Feb 20 09:01:06 crc kubenswrapper[4948]: I0220 09:01:06.579119 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_4b04e19d-cd3d-474d-884d-df12d0e1fffd/openstack-network-exporter/0.log" Feb 20 09:01:06 crc kubenswrapper[4948]: I0220 09:01:06.715503 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_4b04e19d-cd3d-474d-884d-df12d0e1fffd/ovsdbserver-sb/0.log" Feb 20 09:01:06 crc kubenswrapper[4948]: I0220 09:01:06.785647 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-86887846d-pqjcb_5047ff00-845b-459f-ac89-6cbf997f93fb/placement-api/0.log" Feb 20 09:01:06 crc kubenswrapper[4948]: I0220 09:01:06.863795 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-86887846d-pqjcb_5047ff00-845b-459f-ac89-6cbf997f93fb/placement-log/0.log" Feb 20 09:01:06 crc kubenswrapper[4948]: I0220 09:01:06.921845 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_037b329f-712f-4a67-984a-75affd2a57b7/setup-container/0.log" Feb 20 09:01:07 crc kubenswrapper[4948]: I0220 09:01:07.154306 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_037b329f-712f-4a67-984a-75affd2a57b7/setup-container/0.log" Feb 20 09:01:07 crc kubenswrapper[4948]: I0220 09:01:07.178654 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_037b329f-712f-4a67-984a-75affd2a57b7/rabbitmq/0.log" Feb 20 09:01:07 crc kubenswrapper[4948]: I0220 09:01:07.319041 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f063bb62-10c3-46d9-a41d-56d2826059e0/setup-container/0.log" Feb 20 09:01:07 crc kubenswrapper[4948]: I0220 09:01:07.475576 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f063bb62-10c3-46d9-a41d-56d2826059e0/setup-container/0.log" Feb 20 09:01:07 crc kubenswrapper[4948]: I0220 09:01:07.520684 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f063bb62-10c3-46d9-a41d-56d2826059e0/rabbitmq/0.log" Feb 20 09:01:07 crc kubenswrapper[4948]: I0220 09:01:07.561577 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-ds522_e3b05342-f970-4fdb-9b1c-ff521e118f96/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:01:07 crc kubenswrapper[4948]: I0220 09:01:07.758141 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-f7ww9_c3e7cce4-5144-4190-8013-83428c66cde9/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:01:07 crc kubenswrapper[4948]: I0220 09:01:07.787745 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch_90b6e7df-b3e1-4d05-a6b3-7fe47d70804f/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:01:07 crc kubenswrapper[4948]: I0220 09:01:07.973877 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-lkbkf_79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:01:08 crc kubenswrapper[4948]: I0220 09:01:08.024316 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:01:08 crc kubenswrapper[4948]: I0220 09:01:08.024369 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:01:08 crc kubenswrapper[4948]: I0220 09:01:08.042922 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-cqchc_41856163-b621-439a-95c1-ca8d13f0a08c/ssh-known-hosts-edpm-deployment/0.log" Feb 20 09:01:08 crc kubenswrapper[4948]: I0220 09:01:08.277957 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7c7cd9876c-262qf_30f48966-e44d-44bc-a868-7e8119ac186c/proxy-server/0.log" Feb 20 09:01:08 crc kubenswrapper[4948]: I0220 09:01:08.281334 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7c7cd9876c-262qf_30f48966-e44d-44bc-a868-7e8119ac186c/proxy-httpd/0.log" Feb 20 09:01:08 crc kubenswrapper[4948]: I0220 09:01:08.410903 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-dllls_642adfd1-e253-4e0a-80e6-860a9f5c1ae1/swift-ring-rebalance/0.log" Feb 20 09:01:08 crc kubenswrapper[4948]: I0220 09:01:08.516230 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/account-auditor/0.log" Feb 20 09:01:08 crc kubenswrapper[4948]: I0220 09:01:08.596597 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/account-reaper/0.log" Feb 20 09:01:08 crc kubenswrapper[4948]: I0220 09:01:08.679164 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/account-replicator/0.log" Feb 20 09:01:08 crc kubenswrapper[4948]: I0220 09:01:08.698267 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/account-server/0.log" Feb 20 09:01:08 crc kubenswrapper[4948]: I0220 09:01:08.755753 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/container-auditor/0.log" Feb 20 09:01:08 crc kubenswrapper[4948]: I0220 09:01:08.849790 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/container-replicator/0.log" Feb 20 09:01:08 crc kubenswrapper[4948]: I0220 09:01:08.875222 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/container-server/0.log" Feb 20 09:01:08 crc kubenswrapper[4948]: I0220 09:01:08.914707 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/container-updater/0.log" Feb 20 09:01:08 crc kubenswrapper[4948]: I0220 09:01:08.974195 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/object-auditor/0.log" Feb 20 09:01:09 crc kubenswrapper[4948]: I0220 09:01:09.044261 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/object-expirer/0.log" Feb 20 09:01:09 crc kubenswrapper[4948]: I0220 09:01:09.079439 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/object-replicator/0.log" Feb 20 09:01:09 crc kubenswrapper[4948]: I0220 09:01:09.107811 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/object-server/0.log" Feb 20 09:01:09 crc kubenswrapper[4948]: I0220 09:01:09.166741 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/object-updater/0.log" Feb 20 09:01:09 crc kubenswrapper[4948]: I0220 09:01:09.265192 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/rsync/0.log" Feb 20 09:01:09 crc kubenswrapper[4948]: I0220 09:01:09.344093 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/swift-recon-cron/0.log" Feb 20 09:01:09 crc kubenswrapper[4948]: I0220 09:01:09.478441 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd_06ed5839-bc31-4691-8586-cd42c0413006/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:01:09 crc kubenswrapper[4948]: I0220 09:01:09.580106 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_0fefa0c4-ece9-4caf-8b41-8a64472c53e6/tempest-tests-tempest-tests-runner/0.log" Feb 20 09:01:09 crc kubenswrapper[4948]: I0220 09:01:09.699114 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_de083766-efe0-42cd-95c5-b87bbeb33b0a/test-operator-logs-container/0.log" Feb 20 09:01:09 crc kubenswrapper[4948]: I0220 09:01:09.825959 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-lf98d_2e05be38-f4ab-415f-a71d-9e5233cf0ea7/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:01:13 crc kubenswrapper[4948]: I0220 09:01:13.967312 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-7c7cd9876c-262qf" podUID="30f48966-e44d-44bc-a868-7e8119ac186c" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 502" Feb 20 09:01:19 crc kubenswrapper[4948]: I0220 09:01:19.084404 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_8e994489-a2cf-4f6d-a00c-98f627ba0e5f/memcached/0.log" Feb 20 09:01:35 crc kubenswrapper[4948]: I0220 09:01:35.013468 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6_a1b99eb4-2ab2-4755-92ed-a5226d91bfaf/util/0.log" Feb 20 09:01:35 crc kubenswrapper[4948]: I0220 09:01:35.212052 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6_a1b99eb4-2ab2-4755-92ed-a5226d91bfaf/pull/0.log" Feb 20 09:01:35 crc kubenswrapper[4948]: I0220 09:01:35.268492 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6_a1b99eb4-2ab2-4755-92ed-a5226d91bfaf/pull/0.log" Feb 20 09:01:35 crc kubenswrapper[4948]: I0220 09:01:35.271947 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6_a1b99eb4-2ab2-4755-92ed-a5226d91bfaf/util/0.log" Feb 20 09:01:35 crc kubenswrapper[4948]: I0220 09:01:35.472118 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6_a1b99eb4-2ab2-4755-92ed-a5226d91bfaf/pull/0.log" Feb 20 09:01:35 crc kubenswrapper[4948]: I0220 09:01:35.489067 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6_a1b99eb4-2ab2-4755-92ed-a5226d91bfaf/extract/0.log" Feb 20 09:01:35 crc kubenswrapper[4948]: I0220 09:01:35.494681 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6_a1b99eb4-2ab2-4755-92ed-a5226d91bfaf/util/0.log" Feb 20 09:01:35 crc kubenswrapper[4948]: I0220 09:01:35.872628 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-6d8bf5c495-kjpkr_98a1aa06-948b-4034-bc07-7e546e341a8f/manager/0.log" Feb 20 09:01:36 crc kubenswrapper[4948]: I0220 09:01:36.234317 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987464f4-xc74s_347cdd7c-e5e8-49b1-a4f1-687b8a06b250/manager/0.log" Feb 20 09:01:36 crc kubenswrapper[4948]: I0220 09:01:36.370850 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-69f49c598c-6cr5p_2d303d1a-1062-4bf6-be68-1bd6d1a3228f/manager/0.log" Feb 20 09:01:36 crc kubenswrapper[4948]: I0220 09:01:36.585072 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5b9b8895d5-bkl49_4a663826-d4b5-4ed3-8270-099b003390b4/manager/0.log" Feb 20 09:01:36 crc kubenswrapper[4948]: I0220 09:01:36.939601 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-5d946d989d-c6xqd_586113f2-38b6-4bd1-8adb-3c155bb35ba9/manager/0.log" Feb 20 09:01:37 crc kubenswrapper[4948]: I0220 09:01:37.080072 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-554564d7fc-5fqzw_105a118c-121d-4582-960f-1da9957980cb/manager/0.log" Feb 20 09:01:37 crc kubenswrapper[4948]: I0220 09:01:37.120131 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-79d975b745-jrlw5_93b93501-acfe-4274-9a22-ca644b1d11d3/manager/0.log" Feb 20 09:01:37 crc kubenswrapper[4948]: I0220 09:01:37.344521 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-54f6768c69-lj2sz_f762ec0d-e09c-4baf-9540-f4ac61ce7234/manager/0.log" Feb 20 09:01:37 crc kubenswrapper[4948]: I0220 09:01:37.366364 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-b4d948c87-ttkfm_3fbbd11d-612d-479a-a34f-505d995a4871/manager/0.log" Feb 20 09:01:37 crc kubenswrapper[4948]: I0220 09:01:37.613620 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6994f66f48-zcccd_6e8e8b87-5a00-486f-a00c-1450c109f3b2/manager/0.log" Feb 20 09:01:37 crc kubenswrapper[4948]: I0220 09:01:37.875163 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-64ddbf8bb-jcnxd_88142137-864d-4660-a688-a7dcc503851b/manager/0.log" Feb 20 09:01:37 crc kubenswrapper[4948]: I0220 09:01:37.927764 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-567668f5cf-jxwsx_7c7bb531-7900-4cc6-9d9b-bae52dabc59d/manager/0.log" Feb 20 09:01:38 crc kubenswrapper[4948]: I0220 09:01:38.024302 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:01:38 crc kubenswrapper[4948]: I0220 09:01:38.024350 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:01:38 crc kubenswrapper[4948]: I0220 09:01:38.302866 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m_b8b07685-095c-4bbd-a30d-57ca59d7cbdc/manager/0.log" Feb 20 09:01:38 crc kubenswrapper[4948]: I0220 09:01:38.675483 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-59c446d8d6-4cs2n_9465a7a8-c08e-44c9-a76d-3878b0bff3af/operator/0.log" Feb 20 09:01:39 crc kubenswrapper[4948]: I0220 09:01:39.054817 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-9z4jj_64872326-fb95-4bc1-a6c5-0b34242883f9/registry-server/0.log" Feb 20 09:01:39 crc kubenswrapper[4948]: I0220 09:01:39.368806 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-d44cf6b75-dkkq9_3ea675de-d1b0-4880-9652-eb066f6b0fb7/manager/0.log" Feb 20 09:01:39 crc kubenswrapper[4948]: I0220 09:01:39.549209 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69f8888797-hl5w7_ee33bd02-a3ce-415d-9d2d-fefd383e9810/manager/0.log" Feb 20 09:01:39 crc kubenswrapper[4948]: I0220 09:01:39.592378 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-8497b45c89-znrkz_16933747-642c-45ff-9f98-9321c633826a/manager/0.log" Feb 20 09:01:39 crc kubenswrapper[4948]: I0220 09:01:39.763480 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-59pcs_fe0aca33-1acf-463a-91a6-bbf35a38fd7c/operator/0.log" Feb 20 09:01:39 crc kubenswrapper[4948]: I0220 09:01:39.840132 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-68f46476f-5lfts_d65e0993-11a0-4e81-963f-eeb9dcb92536/manager/0.log" Feb 20 09:01:40 crc kubenswrapper[4948]: I0220 09:01:40.051835 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7f45b4ff68-5rrd9_0f0cbb65-3c8e-41e6-8059-34e121de0821/manager/0.log" Feb 20 09:01:40 crc kubenswrapper[4948]: I0220 09:01:40.133756 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-7866795846-f7cxn_0f98bc58-d0e3-405b-88fd-d8bd65f415a4/manager/0.log" Feb 20 09:01:40 crc kubenswrapper[4948]: I0220 09:01:40.287257 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-5db88f68c-pndq7_f2400173-aa87-476c-8216-4f8c9cf9d474/manager/0.log" Feb 20 09:01:40 crc kubenswrapper[4948]: I0220 09:01:40.365532 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-8569bc6fdb-lt4jk_e9f13bcd-ac5a-4cbd-952a-a5b9dceee562/manager/0.log" Feb 20 09:01:42 crc kubenswrapper[4948]: I0220 09:01:42.216770 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-868647ff47-4h9v5_b8e12d0c-9564-4bf2-ac61-b22d2fbdf855/manager/0.log" Feb 20 09:01:59 crc kubenswrapper[4948]: I0220 09:01:59.487175 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-bgm85_a149fe37-c748-4120-9116-1da4b680d880/control-plane-machine-set-operator/0.log" Feb 20 09:01:59 crc kubenswrapper[4948]: I0220 09:01:59.686599 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-fthdq_5bc3d0cd-eacd-4b25-9acd-853e49db7b47/kube-rbac-proxy/0.log" Feb 20 09:01:59 crc kubenswrapper[4948]: I0220 09:01:59.724991 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-fthdq_5bc3d0cd-eacd-4b25-9acd-853e49db7b47/machine-api-operator/0.log" Feb 20 09:02:08 crc kubenswrapper[4948]: I0220 09:02:08.024927 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:02:08 crc kubenswrapper[4948]: I0220 09:02:08.027230 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:02:08 crc kubenswrapper[4948]: I0220 09:02:08.027371 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 09:02:08 crc kubenswrapper[4948]: I0220 09:02:08.028247 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b0e4da4a6a513a5c47c1f0dd87dcf0b56d3fede7e9008d839ac5623f61075ff8"} pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 09:02:08 crc kubenswrapper[4948]: I0220 09:02:08.028409 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" containerID="cri-o://b0e4da4a6a513a5c47c1f0dd87dcf0b56d3fede7e9008d839ac5623f61075ff8" gracePeriod=600 Feb 20 09:02:08 crc kubenswrapper[4948]: I0220 09:02:08.547624 4948 generic.go:334] "Generic (PLEG): container finished" podID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerID="b0e4da4a6a513a5c47c1f0dd87dcf0b56d3fede7e9008d839ac5623f61075ff8" exitCode=0 Feb 20 09:02:08 crc kubenswrapper[4948]: I0220 09:02:08.547808 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerDied","Data":"b0e4da4a6a513a5c47c1f0dd87dcf0b56d3fede7e9008d839ac5623f61075ff8"} Feb 20 09:02:08 crc kubenswrapper[4948]: I0220 09:02:08.548069 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2"} Feb 20 09:02:08 crc kubenswrapper[4948]: I0220 09:02:08.548098 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 09:02:13 crc kubenswrapper[4948]: I0220 09:02:13.471769 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-smtqw_1437d818-abf4-4602-8a28-e88a76e482a9/cert-manager-controller/0.log" Feb 20 09:02:13 crc kubenswrapper[4948]: I0220 09:02:13.679166 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-f4bjc_305a17f4-aef3-4036-8fce-3756ff5bbd2f/cert-manager-webhook/0.log" Feb 20 09:02:13 crc kubenswrapper[4948]: I0220 09:02:13.682806 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-9llfv_bb56399a-a840-49af-972f-4f32c91efed7/cert-manager-cainjector/0.log" Feb 20 09:02:26 crc kubenswrapper[4948]: I0220 09:02:26.888552 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5c78fc5d65-9mv9z_47461ed8-c51b-4014-bcf0-4d95a0278f85/nmstate-console-plugin/0.log" Feb 20 09:02:27 crc kubenswrapper[4948]: I0220 09:02:27.086357 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-dxdh6_3a869b3a-b757-49ba-9096-2e562d980aae/nmstate-handler/0.log" Feb 20 09:02:27 crc kubenswrapper[4948]: I0220 09:02:27.166435 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58c85c668d-tdst7_a7b34c06-d0c5-4f20-9521-eec500ea2510/nmstate-metrics/0.log" Feb 20 09:02:27 crc kubenswrapper[4948]: I0220 09:02:27.179330 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58c85c668d-tdst7_a7b34c06-d0c5-4f20-9521-eec500ea2510/kube-rbac-proxy/0.log" Feb 20 09:02:27 crc kubenswrapper[4948]: I0220 09:02:27.323278 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-694c9596b7-xqtvg_59277281-0112-45da-a64a-1b1d02b90473/nmstate-operator/0.log" Feb 20 09:02:27 crc kubenswrapper[4948]: I0220 09:02:27.357952 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-866bcb46dc-77sdv_da0e4284-f20d-44a8-b529-e388f15822d3/nmstate-webhook/0.log" Feb 20 09:02:56 crc kubenswrapper[4948]: I0220 09:02:56.056653 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-c6sqk_68ee0185-13d4-45c7-9a39-bb474e9314d0/kube-rbac-proxy/0.log" Feb 20 09:02:56 crc kubenswrapper[4948]: I0220 09:02:56.088984 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-c6sqk_68ee0185-13d4-45c7-9a39-bb474e9314d0/controller/0.log" Feb 20 09:02:56 crc kubenswrapper[4948]: I0220 09:02:56.216315 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-frr-files/0.log" Feb 20 09:02:56 crc kubenswrapper[4948]: I0220 09:02:56.431943 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-reloader/0.log" Feb 20 09:02:56 crc kubenswrapper[4948]: I0220 09:02:56.439882 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-reloader/0.log" Feb 20 09:02:56 crc kubenswrapper[4948]: I0220 09:02:56.451433 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-frr-files/0.log" Feb 20 09:02:56 crc kubenswrapper[4948]: I0220 09:02:56.472885 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-metrics/0.log" Feb 20 09:02:56 crc kubenswrapper[4948]: I0220 09:02:56.604499 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-metrics/0.log" Feb 20 09:02:56 crc kubenswrapper[4948]: I0220 09:02:56.612797 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-frr-files/0.log" Feb 20 09:02:56 crc kubenswrapper[4948]: I0220 09:02:56.636472 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-reloader/0.log" Feb 20 09:02:56 crc kubenswrapper[4948]: I0220 09:02:56.679892 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-metrics/0.log" Feb 20 09:02:56 crc kubenswrapper[4948]: I0220 09:02:56.795033 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-reloader/0.log" Feb 20 09:02:56 crc kubenswrapper[4948]: I0220 09:02:56.816872 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-frr-files/0.log" Feb 20 09:02:56 crc kubenswrapper[4948]: I0220 09:02:56.822458 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-metrics/0.log" Feb 20 09:02:56 crc kubenswrapper[4948]: I0220 09:02:56.859885 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/controller/0.log" Feb 20 09:02:56 crc kubenswrapper[4948]: I0220 09:02:56.986285 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/frr-metrics/0.log" Feb 20 09:02:57 crc kubenswrapper[4948]: I0220 09:02:57.113910 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/kube-rbac-proxy-frr/0.log" Feb 20 09:02:57 crc kubenswrapper[4948]: I0220 09:02:57.122798 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/kube-rbac-proxy/0.log" Feb 20 09:02:57 crc kubenswrapper[4948]: I0220 09:02:57.234668 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/reloader/0.log" Feb 20 09:02:57 crc kubenswrapper[4948]: I0220 09:02:57.375166 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-78b44bf5bb-7zsxs_2506d4e2-b434-4fe4-970e-7cd14601677d/frr-k8s-webhook-server/0.log" Feb 20 09:02:57 crc kubenswrapper[4948]: I0220 09:02:57.532751 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-8478c6b8cc-v98zs_b4f86a2e-15f9-441c-953b-49f331d4122e/manager/0.log" Feb 20 09:02:57 crc kubenswrapper[4948]: I0220 09:02:57.633246 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5fcd7d546c-kgpxs_07022b13-7dc1-45d1-a626-91792bd3aa90/webhook-server/0.log" Feb 20 09:02:57 crc kubenswrapper[4948]: I0220 09:02:57.832126 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-h7xhj_2c199acd-2fd3-4b9f-b50c-a266191c4777/kube-rbac-proxy/0.log" Feb 20 09:02:58 crc kubenswrapper[4948]: I0220 09:02:58.260698 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-h7xhj_2c199acd-2fd3-4b9f-b50c-a266191c4777/speaker/0.log" Feb 20 09:02:58 crc kubenswrapper[4948]: I0220 09:02:58.307790 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/frr/0.log" Feb 20 09:03:11 crc kubenswrapper[4948]: I0220 09:03:11.971511 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct_9e5e2f81-8ea9-4ddb-b2ab-584c40b15916/util/0.log" Feb 20 09:03:12 crc kubenswrapper[4948]: I0220 09:03:12.177563 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct_9e5e2f81-8ea9-4ddb-b2ab-584c40b15916/util/0.log" Feb 20 09:03:12 crc kubenswrapper[4948]: I0220 09:03:12.363171 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct_9e5e2f81-8ea9-4ddb-b2ab-584c40b15916/pull/0.log" Feb 20 09:03:12 crc kubenswrapper[4948]: I0220 09:03:12.398645 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct_9e5e2f81-8ea9-4ddb-b2ab-584c40b15916/pull/0.log" Feb 20 09:03:12 crc kubenswrapper[4948]: I0220 09:03:12.583427 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct_9e5e2f81-8ea9-4ddb-b2ab-584c40b15916/pull/0.log" Feb 20 09:03:12 crc kubenswrapper[4948]: I0220 09:03:12.606452 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct_9e5e2f81-8ea9-4ddb-b2ab-584c40b15916/extract/0.log" Feb 20 09:03:12 crc kubenswrapper[4948]: I0220 09:03:12.608145 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct_9e5e2f81-8ea9-4ddb-b2ab-584c40b15916/util/0.log" Feb 20 09:03:12 crc kubenswrapper[4948]: I0220 09:03:12.768800 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tt6t6_94e3bb0b-818a-491e-8e1e-a1664f3a9f12/extract-utilities/0.log" Feb 20 09:03:12 crc kubenswrapper[4948]: I0220 09:03:12.937222 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tt6t6_94e3bb0b-818a-491e-8e1e-a1664f3a9f12/extract-content/0.log" Feb 20 09:03:12 crc kubenswrapper[4948]: I0220 09:03:12.964863 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tt6t6_94e3bb0b-818a-491e-8e1e-a1664f3a9f12/extract-utilities/0.log" Feb 20 09:03:12 crc kubenswrapper[4948]: I0220 09:03:12.996850 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tt6t6_94e3bb0b-818a-491e-8e1e-a1664f3a9f12/extract-content/0.log" Feb 20 09:03:13 crc kubenswrapper[4948]: I0220 09:03:13.124701 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tt6t6_94e3bb0b-818a-491e-8e1e-a1664f3a9f12/extract-utilities/0.log" Feb 20 09:03:13 crc kubenswrapper[4948]: I0220 09:03:13.148464 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tt6t6_94e3bb0b-818a-491e-8e1e-a1664f3a9f12/extract-content/0.log" Feb 20 09:03:13 crc kubenswrapper[4948]: I0220 09:03:13.366946 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dl7nj_0e759dc6-36e8-4a97-b994-e657f1b5d540/extract-utilities/0.log" Feb 20 09:03:13 crc kubenswrapper[4948]: I0220 09:03:13.496734 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dl7nj_0e759dc6-36e8-4a97-b994-e657f1b5d540/extract-content/0.log" Feb 20 09:03:13 crc kubenswrapper[4948]: I0220 09:03:13.508333 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tt6t6_94e3bb0b-818a-491e-8e1e-a1664f3a9f12/registry-server/0.log" Feb 20 09:03:13 crc kubenswrapper[4948]: I0220 09:03:13.529591 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dl7nj_0e759dc6-36e8-4a97-b994-e657f1b5d540/extract-utilities/0.log" Feb 20 09:03:13 crc kubenswrapper[4948]: I0220 09:03:13.597669 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dl7nj_0e759dc6-36e8-4a97-b994-e657f1b5d540/extract-content/0.log" Feb 20 09:03:13 crc kubenswrapper[4948]: I0220 09:03:13.727565 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dl7nj_0e759dc6-36e8-4a97-b994-e657f1b5d540/extract-utilities/0.log" Feb 20 09:03:13 crc kubenswrapper[4948]: I0220 09:03:13.768493 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dl7nj_0e759dc6-36e8-4a97-b994-e657f1b5d540/extract-content/0.log" Feb 20 09:03:13 crc kubenswrapper[4948]: I0220 09:03:13.972889 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w_bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a/util/0.log" Feb 20 09:03:14 crc kubenswrapper[4948]: I0220 09:03:14.145133 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w_bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a/util/0.log" Feb 20 09:03:14 crc kubenswrapper[4948]: I0220 09:03:14.151209 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w_bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a/pull/0.log" Feb 20 09:03:14 crc kubenswrapper[4948]: I0220 09:03:14.249882 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dl7nj_0e759dc6-36e8-4a97-b994-e657f1b5d540/registry-server/0.log" Feb 20 09:03:14 crc kubenswrapper[4948]: I0220 09:03:14.258394 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w_bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a/pull/0.log" Feb 20 09:03:14 crc kubenswrapper[4948]: I0220 09:03:14.440170 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w_bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a/pull/0.log" Feb 20 09:03:14 crc kubenswrapper[4948]: I0220 09:03:14.454049 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w_bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a/util/0.log" Feb 20 09:03:14 crc kubenswrapper[4948]: I0220 09:03:14.455472 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w_bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a/extract/0.log" Feb 20 09:03:14 crc kubenswrapper[4948]: I0220 09:03:14.647526 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x7bcm_6ec98112-e189-4a27-a6dd-8cb035596810/extract-utilities/0.log" Feb 20 09:03:14 crc kubenswrapper[4948]: I0220 09:03:14.721924 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-tbqzs_1e3d2cd4-4311-4454-b257-20a3caf243b3/marketplace-operator/0.log" Feb 20 09:03:14 crc kubenswrapper[4948]: I0220 09:03:14.859745 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x7bcm_6ec98112-e189-4a27-a6dd-8cb035596810/extract-content/0.log" Feb 20 09:03:14 crc kubenswrapper[4948]: I0220 09:03:14.902903 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x7bcm_6ec98112-e189-4a27-a6dd-8cb035596810/extract-utilities/0.log" Feb 20 09:03:14 crc kubenswrapper[4948]: I0220 09:03:14.917430 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x7bcm_6ec98112-e189-4a27-a6dd-8cb035596810/extract-content/0.log" Feb 20 09:03:15 crc kubenswrapper[4948]: I0220 09:03:15.074691 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x7bcm_6ec98112-e189-4a27-a6dd-8cb035596810/extract-utilities/0.log" Feb 20 09:03:15 crc kubenswrapper[4948]: I0220 09:03:15.077897 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x7bcm_6ec98112-e189-4a27-a6dd-8cb035596810/extract-content/0.log" Feb 20 09:03:15 crc kubenswrapper[4948]: I0220 09:03:15.264540 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x7bcm_6ec98112-e189-4a27-a6dd-8cb035596810/registry-server/0.log" Feb 20 09:03:15 crc kubenswrapper[4948]: I0220 09:03:15.311456 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jbtm5_5ee242c9-9783-44cb-a4dc-1ea50a7ae185/extract-utilities/0.log" Feb 20 09:03:15 crc kubenswrapper[4948]: I0220 09:03:15.473484 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jbtm5_5ee242c9-9783-44cb-a4dc-1ea50a7ae185/extract-content/0.log" Feb 20 09:03:15 crc kubenswrapper[4948]: I0220 09:03:15.498000 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jbtm5_5ee242c9-9783-44cb-a4dc-1ea50a7ae185/extract-content/0.log" Feb 20 09:03:15 crc kubenswrapper[4948]: I0220 09:03:15.505267 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jbtm5_5ee242c9-9783-44cb-a4dc-1ea50a7ae185/extract-utilities/0.log" Feb 20 09:03:15 crc kubenswrapper[4948]: I0220 09:03:15.646567 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jbtm5_5ee242c9-9783-44cb-a4dc-1ea50a7ae185/extract-utilities/0.log" Feb 20 09:03:15 crc kubenswrapper[4948]: I0220 09:03:15.711747 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jbtm5_5ee242c9-9783-44cb-a4dc-1ea50a7ae185/extract-content/0.log" Feb 20 09:03:16 crc kubenswrapper[4948]: I0220 09:03:16.027306 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jbtm5_5ee242c9-9783-44cb-a4dc-1ea50a7ae185/registry-server/0.log" Feb 20 09:03:37 crc kubenswrapper[4948]: E0220 09:03:37.505654 4948 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.113:43138->38.102.83.113:38489: write tcp 38.102.83.113:43138->38.102.83.113:38489: write: broken pipe Feb 20 09:04:08 crc kubenswrapper[4948]: I0220 09:04:08.024915 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:04:08 crc kubenswrapper[4948]: I0220 09:04:08.025561 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:04:34 crc kubenswrapper[4948]: I0220 09:04:34.337836 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cqvgj"] Feb 20 09:04:34 crc kubenswrapper[4948]: E0220 09:04:34.339106 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7990f62f-cd0f-4be2-b166-c00621d877e1" containerName="keystone-cron" Feb 20 09:04:34 crc kubenswrapper[4948]: I0220 09:04:34.339132 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="7990f62f-cd0f-4be2-b166-c00621d877e1" containerName="keystone-cron" Feb 20 09:04:34 crc kubenswrapper[4948]: I0220 09:04:34.339421 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="7990f62f-cd0f-4be2-b166-c00621d877e1" containerName="keystone-cron" Feb 20 09:04:34 crc kubenswrapper[4948]: I0220 09:04:34.341522 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cqvgj" Feb 20 09:04:34 crc kubenswrapper[4948]: I0220 09:04:34.352287 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cqvgj"] Feb 20 09:04:34 crc kubenswrapper[4948]: I0220 09:04:34.413804 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-utilities\") pod \"redhat-operators-cqvgj\" (UID: \"fa0586ff-df65-4d4d-8571-eb1d0bed5b62\") " pod="openshift-marketplace/redhat-operators-cqvgj" Feb 20 09:04:34 crc kubenswrapper[4948]: I0220 09:04:34.413860 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lphb2\" (UniqueName: \"kubernetes.io/projected/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-kube-api-access-lphb2\") pod \"redhat-operators-cqvgj\" (UID: \"fa0586ff-df65-4d4d-8571-eb1d0bed5b62\") " pod="openshift-marketplace/redhat-operators-cqvgj" Feb 20 09:04:34 crc kubenswrapper[4948]: I0220 09:04:34.413940 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-catalog-content\") pod \"redhat-operators-cqvgj\" (UID: \"fa0586ff-df65-4d4d-8571-eb1d0bed5b62\") " pod="openshift-marketplace/redhat-operators-cqvgj" Feb 20 09:04:34 crc kubenswrapper[4948]: I0220 09:04:34.515743 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-utilities\") pod \"redhat-operators-cqvgj\" (UID: \"fa0586ff-df65-4d4d-8571-eb1d0bed5b62\") " pod="openshift-marketplace/redhat-operators-cqvgj" Feb 20 09:04:34 crc kubenswrapper[4948]: I0220 09:04:34.515782 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lphb2\" (UniqueName: \"kubernetes.io/projected/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-kube-api-access-lphb2\") pod \"redhat-operators-cqvgj\" (UID: \"fa0586ff-df65-4d4d-8571-eb1d0bed5b62\") " pod="openshift-marketplace/redhat-operators-cqvgj" Feb 20 09:04:34 crc kubenswrapper[4948]: I0220 09:04:34.515843 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-catalog-content\") pod \"redhat-operators-cqvgj\" (UID: \"fa0586ff-df65-4d4d-8571-eb1d0bed5b62\") " pod="openshift-marketplace/redhat-operators-cqvgj" Feb 20 09:04:34 crc kubenswrapper[4948]: I0220 09:04:34.516350 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-utilities\") pod \"redhat-operators-cqvgj\" (UID: \"fa0586ff-df65-4d4d-8571-eb1d0bed5b62\") " pod="openshift-marketplace/redhat-operators-cqvgj" Feb 20 09:04:34 crc kubenswrapper[4948]: I0220 09:04:34.516416 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-catalog-content\") pod \"redhat-operators-cqvgj\" (UID: \"fa0586ff-df65-4d4d-8571-eb1d0bed5b62\") " pod="openshift-marketplace/redhat-operators-cqvgj" Feb 20 09:04:34 crc kubenswrapper[4948]: I0220 09:04:34.539156 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lphb2\" (UniqueName: \"kubernetes.io/projected/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-kube-api-access-lphb2\") pod \"redhat-operators-cqvgj\" (UID: \"fa0586ff-df65-4d4d-8571-eb1d0bed5b62\") " pod="openshift-marketplace/redhat-operators-cqvgj" Feb 20 09:04:34 crc kubenswrapper[4948]: I0220 09:04:34.688777 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cqvgj" Feb 20 09:04:35 crc kubenswrapper[4948]: I0220 09:04:35.166396 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cqvgj"] Feb 20 09:04:36 crc kubenswrapper[4948]: I0220 09:04:36.010062 4948 generic.go:334] "Generic (PLEG): container finished" podID="fa0586ff-df65-4d4d-8571-eb1d0bed5b62" containerID="d54544788f251e4b6d29e4c674bc8416aca134867b24b018c4f49416d726ab6e" exitCode=0 Feb 20 09:04:36 crc kubenswrapper[4948]: I0220 09:04:36.010297 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cqvgj" event={"ID":"fa0586ff-df65-4d4d-8571-eb1d0bed5b62","Type":"ContainerDied","Data":"d54544788f251e4b6d29e4c674bc8416aca134867b24b018c4f49416d726ab6e"} Feb 20 09:04:36 crc kubenswrapper[4948]: I0220 09:04:36.010389 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cqvgj" event={"ID":"fa0586ff-df65-4d4d-8571-eb1d0bed5b62","Type":"ContainerStarted","Data":"e0a78829b49373d7761eae4a5b21ce062c4d2de49690ef10c9018d2c5d7916c2"} Feb 20 09:04:36 crc kubenswrapper[4948]: I0220 09:04:36.018150 4948 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 09:04:37 crc kubenswrapper[4948]: I0220 09:04:37.022128 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cqvgj" event={"ID":"fa0586ff-df65-4d4d-8571-eb1d0bed5b62","Type":"ContainerStarted","Data":"49c59846e9d076d7fd5056bdfdce13873c0177f11109eb4581cf9a83aaa33bdc"} Feb 20 09:04:38 crc kubenswrapper[4948]: I0220 09:04:38.024496 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:04:38 crc kubenswrapper[4948]: I0220 09:04:38.024569 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:04:39 crc kubenswrapper[4948]: I0220 09:04:39.040776 4948 generic.go:334] "Generic (PLEG): container finished" podID="fa0586ff-df65-4d4d-8571-eb1d0bed5b62" containerID="49c59846e9d076d7fd5056bdfdce13873c0177f11109eb4581cf9a83aaa33bdc" exitCode=0 Feb 20 09:04:39 crc kubenswrapper[4948]: I0220 09:04:39.040852 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cqvgj" event={"ID":"fa0586ff-df65-4d4d-8571-eb1d0bed5b62","Type":"ContainerDied","Data":"49c59846e9d076d7fd5056bdfdce13873c0177f11109eb4581cf9a83aaa33bdc"} Feb 20 09:04:40 crc kubenswrapper[4948]: I0220 09:04:40.060766 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cqvgj" event={"ID":"fa0586ff-df65-4d4d-8571-eb1d0bed5b62","Type":"ContainerStarted","Data":"aba938c4a927bdefe6c39509a204ce9987b639f0088daf4bdfdf4018402bc5e8"} Feb 20 09:04:40 crc kubenswrapper[4948]: I0220 09:04:40.097828 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cqvgj" podStartSLOduration=2.593158372 podStartE2EDuration="6.097802944s" podCreationTimestamp="2026-02-20 09:04:34 +0000 UTC" firstStartedPulling="2026-02-20 09:04:36.017915291 +0000 UTC m=+3524.992410111" lastFinishedPulling="2026-02-20 09:04:39.522559853 +0000 UTC m=+3528.497054683" observedRunningTime="2026-02-20 09:04:40.089275413 +0000 UTC m=+3529.063770273" watchObservedRunningTime="2026-02-20 09:04:40.097802944 +0000 UTC m=+3529.072297794" Feb 20 09:04:44 crc kubenswrapper[4948]: I0220 09:04:44.689554 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cqvgj" Feb 20 09:04:44 crc kubenswrapper[4948]: I0220 09:04:44.691838 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cqvgj" Feb 20 09:04:45 crc kubenswrapper[4948]: I0220 09:04:45.772575 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-cqvgj" podUID="fa0586ff-df65-4d4d-8571-eb1d0bed5b62" containerName="registry-server" probeResult="failure" output=< Feb 20 09:04:45 crc kubenswrapper[4948]: timeout: failed to connect service ":50051" within 1s Feb 20 09:04:45 crc kubenswrapper[4948]: > Feb 20 09:04:53 crc kubenswrapper[4948]: I0220 09:04:53.102549 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2hl84"] Feb 20 09:04:53 crc kubenswrapper[4948]: I0220 09:04:53.108255 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2hl84" Feb 20 09:04:53 crc kubenswrapper[4948]: I0220 09:04:53.124436 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2hl84"] Feb 20 09:04:53 crc kubenswrapper[4948]: I0220 09:04:53.281057 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hv8x\" (UniqueName: \"kubernetes.io/projected/27f74a12-0843-4c61-a5c5-7f026224d86e-kube-api-access-6hv8x\") pod \"certified-operators-2hl84\" (UID: \"27f74a12-0843-4c61-a5c5-7f026224d86e\") " pod="openshift-marketplace/certified-operators-2hl84" Feb 20 09:04:53 crc kubenswrapper[4948]: I0220 09:04:53.281145 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27f74a12-0843-4c61-a5c5-7f026224d86e-catalog-content\") pod \"certified-operators-2hl84\" (UID: \"27f74a12-0843-4c61-a5c5-7f026224d86e\") " pod="openshift-marketplace/certified-operators-2hl84" Feb 20 09:04:53 crc kubenswrapper[4948]: I0220 09:04:53.281164 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27f74a12-0843-4c61-a5c5-7f026224d86e-utilities\") pod \"certified-operators-2hl84\" (UID: \"27f74a12-0843-4c61-a5c5-7f026224d86e\") " pod="openshift-marketplace/certified-operators-2hl84" Feb 20 09:04:53 crc kubenswrapper[4948]: I0220 09:04:53.383197 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27f74a12-0843-4c61-a5c5-7f026224d86e-catalog-content\") pod \"certified-operators-2hl84\" (UID: \"27f74a12-0843-4c61-a5c5-7f026224d86e\") " pod="openshift-marketplace/certified-operators-2hl84" Feb 20 09:04:53 crc kubenswrapper[4948]: I0220 09:04:53.383243 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27f74a12-0843-4c61-a5c5-7f026224d86e-utilities\") pod \"certified-operators-2hl84\" (UID: \"27f74a12-0843-4c61-a5c5-7f026224d86e\") " pod="openshift-marketplace/certified-operators-2hl84" Feb 20 09:04:53 crc kubenswrapper[4948]: I0220 09:04:53.383370 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hv8x\" (UniqueName: \"kubernetes.io/projected/27f74a12-0843-4c61-a5c5-7f026224d86e-kube-api-access-6hv8x\") pod \"certified-operators-2hl84\" (UID: \"27f74a12-0843-4c61-a5c5-7f026224d86e\") " pod="openshift-marketplace/certified-operators-2hl84" Feb 20 09:04:53 crc kubenswrapper[4948]: I0220 09:04:53.383671 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27f74a12-0843-4c61-a5c5-7f026224d86e-catalog-content\") pod \"certified-operators-2hl84\" (UID: \"27f74a12-0843-4c61-a5c5-7f026224d86e\") " pod="openshift-marketplace/certified-operators-2hl84" Feb 20 09:04:53 crc kubenswrapper[4948]: I0220 09:04:53.384053 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27f74a12-0843-4c61-a5c5-7f026224d86e-utilities\") pod \"certified-operators-2hl84\" (UID: \"27f74a12-0843-4c61-a5c5-7f026224d86e\") " pod="openshift-marketplace/certified-operators-2hl84" Feb 20 09:04:53 crc kubenswrapper[4948]: I0220 09:04:53.402166 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hv8x\" (UniqueName: \"kubernetes.io/projected/27f74a12-0843-4c61-a5c5-7f026224d86e-kube-api-access-6hv8x\") pod \"certified-operators-2hl84\" (UID: \"27f74a12-0843-4c61-a5c5-7f026224d86e\") " pod="openshift-marketplace/certified-operators-2hl84" Feb 20 09:04:53 crc kubenswrapper[4948]: I0220 09:04:53.473022 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2hl84" Feb 20 09:04:54 crc kubenswrapper[4948]: I0220 09:04:54.026826 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2hl84"] Feb 20 09:04:54 crc kubenswrapper[4948]: I0220 09:04:54.199499 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2hl84" event={"ID":"27f74a12-0843-4c61-a5c5-7f026224d86e","Type":"ContainerStarted","Data":"9d10453c1bde550a3b492982c9301c05c35c72c85021ea721f6d12fcfa8a9df5"} Feb 20 09:04:54 crc kubenswrapper[4948]: I0220 09:04:54.789935 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cqvgj" Feb 20 09:04:54 crc kubenswrapper[4948]: I0220 09:04:54.854201 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cqvgj" Feb 20 09:04:55 crc kubenswrapper[4948]: I0220 09:04:55.215257 4948 generic.go:334] "Generic (PLEG): container finished" podID="27f74a12-0843-4c61-a5c5-7f026224d86e" containerID="274f49b6cdbfb4b1df5dfe4a37883e06663d4ef43a70cc2e8f1bc96bbdbcfa76" exitCode=0 Feb 20 09:04:55 crc kubenswrapper[4948]: I0220 09:04:55.215334 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2hl84" event={"ID":"27f74a12-0843-4c61-a5c5-7f026224d86e","Type":"ContainerDied","Data":"274f49b6cdbfb4b1df5dfe4a37883e06663d4ef43a70cc2e8f1bc96bbdbcfa76"} Feb 20 09:04:57 crc kubenswrapper[4948]: I0220 09:04:57.084678 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cqvgj"] Feb 20 09:04:57 crc kubenswrapper[4948]: I0220 09:04:57.085454 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cqvgj" podUID="fa0586ff-df65-4d4d-8571-eb1d0bed5b62" containerName="registry-server" containerID="cri-o://aba938c4a927bdefe6c39509a204ce9987b639f0088daf4bdfdf4018402bc5e8" gracePeriod=2 Feb 20 09:04:57 crc kubenswrapper[4948]: I0220 09:04:57.234362 4948 generic.go:334] "Generic (PLEG): container finished" podID="27f74a12-0843-4c61-a5c5-7f026224d86e" containerID="b940cc80e090937cbff77b5ed4366ab50ebcea5b7644547c7ab3fa22e20c4eeb" exitCode=0 Feb 20 09:04:57 crc kubenswrapper[4948]: I0220 09:04:57.234433 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2hl84" event={"ID":"27f74a12-0843-4c61-a5c5-7f026224d86e","Type":"ContainerDied","Data":"b940cc80e090937cbff77b5ed4366ab50ebcea5b7644547c7ab3fa22e20c4eeb"} Feb 20 09:04:57 crc kubenswrapper[4948]: I0220 09:04:57.243071 4948 generic.go:334] "Generic (PLEG): container finished" podID="fa0586ff-df65-4d4d-8571-eb1d0bed5b62" containerID="aba938c4a927bdefe6c39509a204ce9987b639f0088daf4bdfdf4018402bc5e8" exitCode=0 Feb 20 09:04:57 crc kubenswrapper[4948]: I0220 09:04:57.243116 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cqvgj" event={"ID":"fa0586ff-df65-4d4d-8571-eb1d0bed5b62","Type":"ContainerDied","Data":"aba938c4a927bdefe6c39509a204ce9987b639f0088daf4bdfdf4018402bc5e8"} Feb 20 09:04:57 crc kubenswrapper[4948]: I0220 09:04:57.583487 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cqvgj" Feb 20 09:04:57 crc kubenswrapper[4948]: I0220 09:04:57.679400 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-utilities\") pod \"fa0586ff-df65-4d4d-8571-eb1d0bed5b62\" (UID: \"fa0586ff-df65-4d4d-8571-eb1d0bed5b62\") " Feb 20 09:04:57 crc kubenswrapper[4948]: I0220 09:04:57.679501 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lphb2\" (UniqueName: \"kubernetes.io/projected/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-kube-api-access-lphb2\") pod \"fa0586ff-df65-4d4d-8571-eb1d0bed5b62\" (UID: \"fa0586ff-df65-4d4d-8571-eb1d0bed5b62\") " Feb 20 09:04:57 crc kubenswrapper[4948]: I0220 09:04:57.679674 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-catalog-content\") pod \"fa0586ff-df65-4d4d-8571-eb1d0bed5b62\" (UID: \"fa0586ff-df65-4d4d-8571-eb1d0bed5b62\") " Feb 20 09:04:57 crc kubenswrapper[4948]: I0220 09:04:57.690878 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-utilities" (OuterVolumeSpecName: "utilities") pod "fa0586ff-df65-4d4d-8571-eb1d0bed5b62" (UID: "fa0586ff-df65-4d4d-8571-eb1d0bed5b62"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:04:57 crc kubenswrapper[4948]: I0220 09:04:57.701632 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-kube-api-access-lphb2" (OuterVolumeSpecName: "kube-api-access-lphb2") pod "fa0586ff-df65-4d4d-8571-eb1d0bed5b62" (UID: "fa0586ff-df65-4d4d-8571-eb1d0bed5b62"). InnerVolumeSpecName "kube-api-access-lphb2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:04:57 crc kubenswrapper[4948]: I0220 09:04:57.782274 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lphb2\" (UniqueName: \"kubernetes.io/projected/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-kube-api-access-lphb2\") on node \"crc\" DevicePath \"\"" Feb 20 09:04:57 crc kubenswrapper[4948]: I0220 09:04:57.782313 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 09:04:57 crc kubenswrapper[4948]: I0220 09:04:57.837251 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fa0586ff-df65-4d4d-8571-eb1d0bed5b62" (UID: "fa0586ff-df65-4d4d-8571-eb1d0bed5b62"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:04:57 crc kubenswrapper[4948]: I0220 09:04:57.884927 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 09:04:58 crc kubenswrapper[4948]: I0220 09:04:58.268217 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2hl84" event={"ID":"27f74a12-0843-4c61-a5c5-7f026224d86e","Type":"ContainerStarted","Data":"42844cc459dfe5a86d4c469b9410a53b0ae7589280ea76862abd9dfba05593ca"} Feb 20 09:04:58 crc kubenswrapper[4948]: I0220 09:04:58.271711 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cqvgj" event={"ID":"fa0586ff-df65-4d4d-8571-eb1d0bed5b62","Type":"ContainerDied","Data":"e0a78829b49373d7761eae4a5b21ce062c4d2de49690ef10c9018d2c5d7916c2"} Feb 20 09:04:58 crc kubenswrapper[4948]: I0220 09:04:58.271764 4948 scope.go:117] "RemoveContainer" containerID="aba938c4a927bdefe6c39509a204ce9987b639f0088daf4bdfdf4018402bc5e8" Feb 20 09:04:58 crc kubenswrapper[4948]: I0220 09:04:58.271769 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cqvgj" Feb 20 09:04:58 crc kubenswrapper[4948]: I0220 09:04:58.301586 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2hl84" podStartSLOduration=2.870851341 podStartE2EDuration="5.301560437s" podCreationTimestamp="2026-02-20 09:04:53 +0000 UTC" firstStartedPulling="2026-02-20 09:04:55.221587873 +0000 UTC m=+3544.196082743" lastFinishedPulling="2026-02-20 09:04:57.652297019 +0000 UTC m=+3546.626791839" observedRunningTime="2026-02-20 09:04:58.301396043 +0000 UTC m=+3547.275890893" watchObservedRunningTime="2026-02-20 09:04:58.301560437 +0000 UTC m=+3547.276055287" Feb 20 09:04:58 crc kubenswrapper[4948]: I0220 09:04:58.323603 4948 scope.go:117] "RemoveContainer" containerID="49c59846e9d076d7fd5056bdfdce13873c0177f11109eb4581cf9a83aaa33bdc" Feb 20 09:04:58 crc kubenswrapper[4948]: I0220 09:04:58.335867 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cqvgj"] Feb 20 09:04:58 crc kubenswrapper[4948]: I0220 09:04:58.355931 4948 scope.go:117] "RemoveContainer" containerID="d54544788f251e4b6d29e4c674bc8416aca134867b24b018c4f49416d726ab6e" Feb 20 09:04:58 crc kubenswrapper[4948]: I0220 09:04:58.358464 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cqvgj"] Feb 20 09:04:59 crc kubenswrapper[4948]: I0220 09:04:59.737953 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa0586ff-df65-4d4d-8571-eb1d0bed5b62" path="/var/lib/kubelet/pods/fa0586ff-df65-4d4d-8571-eb1d0bed5b62/volumes" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.084830 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gs7f6"] Feb 20 09:05:02 crc kubenswrapper[4948]: E0220 09:05:02.085634 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa0586ff-df65-4d4d-8571-eb1d0bed5b62" containerName="extract-content" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.085648 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa0586ff-df65-4d4d-8571-eb1d0bed5b62" containerName="extract-content" Feb 20 09:05:02 crc kubenswrapper[4948]: E0220 09:05:02.085665 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa0586ff-df65-4d4d-8571-eb1d0bed5b62" containerName="registry-server" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.085671 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa0586ff-df65-4d4d-8571-eb1d0bed5b62" containerName="registry-server" Feb 20 09:05:02 crc kubenswrapper[4948]: E0220 09:05:02.085689 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa0586ff-df65-4d4d-8571-eb1d0bed5b62" containerName="extract-utilities" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.085696 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa0586ff-df65-4d4d-8571-eb1d0bed5b62" containerName="extract-utilities" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.085877 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa0586ff-df65-4d4d-8571-eb1d0bed5b62" containerName="registry-server" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.087467 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gs7f6" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.113569 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gs7f6"] Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.283192 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-catalog-content\") pod \"community-operators-gs7f6\" (UID: \"fb1bc246-7f5a-4060-8bdb-6a3963ea6862\") " pod="openshift-marketplace/community-operators-gs7f6" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.283248 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-utilities\") pod \"community-operators-gs7f6\" (UID: \"fb1bc246-7f5a-4060-8bdb-6a3963ea6862\") " pod="openshift-marketplace/community-operators-gs7f6" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.283457 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmpwq\" (UniqueName: \"kubernetes.io/projected/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-kube-api-access-bmpwq\") pod \"community-operators-gs7f6\" (UID: \"fb1bc246-7f5a-4060-8bdb-6a3963ea6862\") " pod="openshift-marketplace/community-operators-gs7f6" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.322823 4948 generic.go:334] "Generic (PLEG): container finished" podID="b25c36d5-5de9-4c98-acf4-495d492def68" containerID="4a8285325687a203e4a7cd75d1ec55e566e63219bb19f28c257e627288bb520a" exitCode=0 Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.322868 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8sqsd/must-gather-xzw72" event={"ID":"b25c36d5-5de9-4c98-acf4-495d492def68","Type":"ContainerDied","Data":"4a8285325687a203e4a7cd75d1ec55e566e63219bb19f28c257e627288bb520a"} Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.323505 4948 scope.go:117] "RemoveContainer" containerID="4a8285325687a203e4a7cd75d1ec55e566e63219bb19f28c257e627288bb520a" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.385836 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-catalog-content\") pod \"community-operators-gs7f6\" (UID: \"fb1bc246-7f5a-4060-8bdb-6a3963ea6862\") " pod="openshift-marketplace/community-operators-gs7f6" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.385876 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-utilities\") pod \"community-operators-gs7f6\" (UID: \"fb1bc246-7f5a-4060-8bdb-6a3963ea6862\") " pod="openshift-marketplace/community-operators-gs7f6" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.385929 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmpwq\" (UniqueName: \"kubernetes.io/projected/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-kube-api-access-bmpwq\") pod \"community-operators-gs7f6\" (UID: \"fb1bc246-7f5a-4060-8bdb-6a3963ea6862\") " pod="openshift-marketplace/community-operators-gs7f6" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.386517 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-catalog-content\") pod \"community-operators-gs7f6\" (UID: \"fb1bc246-7f5a-4060-8bdb-6a3963ea6862\") " pod="openshift-marketplace/community-operators-gs7f6" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.386579 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-utilities\") pod \"community-operators-gs7f6\" (UID: \"fb1bc246-7f5a-4060-8bdb-6a3963ea6862\") " pod="openshift-marketplace/community-operators-gs7f6" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.407958 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmpwq\" (UniqueName: \"kubernetes.io/projected/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-kube-api-access-bmpwq\") pod \"community-operators-gs7f6\" (UID: \"fb1bc246-7f5a-4060-8bdb-6a3963ea6862\") " pod="openshift-marketplace/community-operators-gs7f6" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.428133 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gs7f6" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.939727 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gs7f6"] Feb 20 09:05:02 crc kubenswrapper[4948]: W0220 09:05:02.950236 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfb1bc246_7f5a_4060_8bdb_6a3963ea6862.slice/crio-a09d366d7a1bea20bacda46f1149340ebdb8bdbac35df3dd60174f0f350bbfa7 WatchSource:0}: Error finding container a09d366d7a1bea20bacda46f1149340ebdb8bdbac35df3dd60174f0f350bbfa7: Status 404 returned error can't find the container with id a09d366d7a1bea20bacda46f1149340ebdb8bdbac35df3dd60174f0f350bbfa7 Feb 20 09:05:03 crc kubenswrapper[4948]: I0220 09:05:03.102552 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-8sqsd_must-gather-xzw72_b25c36d5-5de9-4c98-acf4-495d492def68/gather/0.log" Feb 20 09:05:03 crc kubenswrapper[4948]: I0220 09:05:03.335615 4948 generic.go:334] "Generic (PLEG): container finished" podID="fb1bc246-7f5a-4060-8bdb-6a3963ea6862" containerID="ac1cf50592cbedded2cb785f164d800b487c31663eb504bce695abf3f1e85685" exitCode=0 Feb 20 09:05:03 crc kubenswrapper[4948]: I0220 09:05:03.335657 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gs7f6" event={"ID":"fb1bc246-7f5a-4060-8bdb-6a3963ea6862","Type":"ContainerDied","Data":"ac1cf50592cbedded2cb785f164d800b487c31663eb504bce695abf3f1e85685"} Feb 20 09:05:03 crc kubenswrapper[4948]: I0220 09:05:03.335684 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gs7f6" event={"ID":"fb1bc246-7f5a-4060-8bdb-6a3963ea6862","Type":"ContainerStarted","Data":"a09d366d7a1bea20bacda46f1149340ebdb8bdbac35df3dd60174f0f350bbfa7"} Feb 20 09:05:03 crc kubenswrapper[4948]: I0220 09:05:03.473731 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2hl84" Feb 20 09:05:03 crc kubenswrapper[4948]: I0220 09:05:03.473776 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2hl84" Feb 20 09:05:03 crc kubenswrapper[4948]: I0220 09:05:03.534454 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2hl84" Feb 20 09:05:04 crc kubenswrapper[4948]: I0220 09:05:04.350881 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gs7f6" event={"ID":"fb1bc246-7f5a-4060-8bdb-6a3963ea6862","Type":"ContainerStarted","Data":"d5289ac930f2ff05d94bb857ba6dd563c33915f3a369ab91433c852e800b4d85"} Feb 20 09:05:04 crc kubenswrapper[4948]: I0220 09:05:04.401709 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2hl84" Feb 20 09:05:05 crc kubenswrapper[4948]: I0220 09:05:05.365216 4948 generic.go:334] "Generic (PLEG): container finished" podID="fb1bc246-7f5a-4060-8bdb-6a3963ea6862" containerID="d5289ac930f2ff05d94bb857ba6dd563c33915f3a369ab91433c852e800b4d85" exitCode=0 Feb 20 09:05:05 crc kubenswrapper[4948]: I0220 09:05:05.365270 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gs7f6" event={"ID":"fb1bc246-7f5a-4060-8bdb-6a3963ea6862","Type":"ContainerDied","Data":"d5289ac930f2ff05d94bb857ba6dd563c33915f3a369ab91433c852e800b4d85"} Feb 20 09:05:05 crc kubenswrapper[4948]: I0220 09:05:05.890705 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2hl84"] Feb 20 09:05:06 crc kubenswrapper[4948]: I0220 09:05:06.378999 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gs7f6" event={"ID":"fb1bc246-7f5a-4060-8bdb-6a3963ea6862","Type":"ContainerStarted","Data":"85b52de1041bebd75c7433b32671f81a45c491a3db70365771075aac31cce321"} Feb 20 09:05:06 crc kubenswrapper[4948]: I0220 09:05:06.379163 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2hl84" podUID="27f74a12-0843-4c61-a5c5-7f026224d86e" containerName="registry-server" containerID="cri-o://42844cc459dfe5a86d4c469b9410a53b0ae7589280ea76862abd9dfba05593ca" gracePeriod=2 Feb 20 09:05:06 crc kubenswrapper[4948]: I0220 09:05:06.406371 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gs7f6" podStartSLOduration=1.944916492 podStartE2EDuration="4.406350816s" podCreationTimestamp="2026-02-20 09:05:02 +0000 UTC" firstStartedPulling="2026-02-20 09:05:03.338386708 +0000 UTC m=+3552.312881528" lastFinishedPulling="2026-02-20 09:05:05.799821022 +0000 UTC m=+3554.774315852" observedRunningTime="2026-02-20 09:05:06.403062845 +0000 UTC m=+3555.377557665" watchObservedRunningTime="2026-02-20 09:05:06.406350816 +0000 UTC m=+3555.380845646" Feb 20 09:05:06 crc kubenswrapper[4948]: I0220 09:05:06.852156 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2hl84" Feb 20 09:05:06 crc kubenswrapper[4948]: I0220 09:05:06.985332 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27f74a12-0843-4c61-a5c5-7f026224d86e-catalog-content\") pod \"27f74a12-0843-4c61-a5c5-7f026224d86e\" (UID: \"27f74a12-0843-4c61-a5c5-7f026224d86e\") " Feb 20 09:05:06 crc kubenswrapper[4948]: I0220 09:05:06.985468 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27f74a12-0843-4c61-a5c5-7f026224d86e-utilities\") pod \"27f74a12-0843-4c61-a5c5-7f026224d86e\" (UID: \"27f74a12-0843-4c61-a5c5-7f026224d86e\") " Feb 20 09:05:06 crc kubenswrapper[4948]: I0220 09:05:06.985590 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hv8x\" (UniqueName: \"kubernetes.io/projected/27f74a12-0843-4c61-a5c5-7f026224d86e-kube-api-access-6hv8x\") pod \"27f74a12-0843-4c61-a5c5-7f026224d86e\" (UID: \"27f74a12-0843-4c61-a5c5-7f026224d86e\") " Feb 20 09:05:06 crc kubenswrapper[4948]: I0220 09:05:06.986910 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27f74a12-0843-4c61-a5c5-7f026224d86e-utilities" (OuterVolumeSpecName: "utilities") pod "27f74a12-0843-4c61-a5c5-7f026224d86e" (UID: "27f74a12-0843-4c61-a5c5-7f026224d86e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:05:06 crc kubenswrapper[4948]: I0220 09:05:06.993345 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27f74a12-0843-4c61-a5c5-7f026224d86e-kube-api-access-6hv8x" (OuterVolumeSpecName: "kube-api-access-6hv8x") pod "27f74a12-0843-4c61-a5c5-7f026224d86e" (UID: "27f74a12-0843-4c61-a5c5-7f026224d86e"). InnerVolumeSpecName "kube-api-access-6hv8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.039192 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27f74a12-0843-4c61-a5c5-7f026224d86e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "27f74a12-0843-4c61-a5c5-7f026224d86e" (UID: "27f74a12-0843-4c61-a5c5-7f026224d86e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.087917 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27f74a12-0843-4c61-a5c5-7f026224d86e-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.087958 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27f74a12-0843-4c61-a5c5-7f026224d86e-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.087984 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hv8x\" (UniqueName: \"kubernetes.io/projected/27f74a12-0843-4c61-a5c5-7f026224d86e-kube-api-access-6hv8x\") on node \"crc\" DevicePath \"\"" Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.389891 4948 generic.go:334] "Generic (PLEG): container finished" podID="27f74a12-0843-4c61-a5c5-7f026224d86e" containerID="42844cc459dfe5a86d4c469b9410a53b0ae7589280ea76862abd9dfba05593ca" exitCode=0 Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.389992 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2hl84" event={"ID":"27f74a12-0843-4c61-a5c5-7f026224d86e","Type":"ContainerDied","Data":"42844cc459dfe5a86d4c469b9410a53b0ae7589280ea76862abd9dfba05593ca"} Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.390044 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2hl84" event={"ID":"27f74a12-0843-4c61-a5c5-7f026224d86e","Type":"ContainerDied","Data":"9d10453c1bde550a3b492982c9301c05c35c72c85021ea721f6d12fcfa8a9df5"} Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.390071 4948 scope.go:117] "RemoveContainer" containerID="42844cc459dfe5a86d4c469b9410a53b0ae7589280ea76862abd9dfba05593ca" Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.390230 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2hl84" Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.433138 4948 scope.go:117] "RemoveContainer" containerID="b940cc80e090937cbff77b5ed4366ab50ebcea5b7644547c7ab3fa22e20c4eeb" Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.434456 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2hl84"] Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.444836 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2hl84"] Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.456858 4948 scope.go:117] "RemoveContainer" containerID="274f49b6cdbfb4b1df5dfe4a37883e06663d4ef43a70cc2e8f1bc96bbdbcfa76" Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.496874 4948 scope.go:117] "RemoveContainer" containerID="42844cc459dfe5a86d4c469b9410a53b0ae7589280ea76862abd9dfba05593ca" Feb 20 09:05:07 crc kubenswrapper[4948]: E0220 09:05:07.497320 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42844cc459dfe5a86d4c469b9410a53b0ae7589280ea76862abd9dfba05593ca\": container with ID starting with 42844cc459dfe5a86d4c469b9410a53b0ae7589280ea76862abd9dfba05593ca not found: ID does not exist" containerID="42844cc459dfe5a86d4c469b9410a53b0ae7589280ea76862abd9dfba05593ca" Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.497406 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42844cc459dfe5a86d4c469b9410a53b0ae7589280ea76862abd9dfba05593ca"} err="failed to get container status \"42844cc459dfe5a86d4c469b9410a53b0ae7589280ea76862abd9dfba05593ca\": rpc error: code = NotFound desc = could not find container \"42844cc459dfe5a86d4c469b9410a53b0ae7589280ea76862abd9dfba05593ca\": container with ID starting with 42844cc459dfe5a86d4c469b9410a53b0ae7589280ea76862abd9dfba05593ca not found: ID does not exist" Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.497471 4948 scope.go:117] "RemoveContainer" containerID="b940cc80e090937cbff77b5ed4366ab50ebcea5b7644547c7ab3fa22e20c4eeb" Feb 20 09:05:07 crc kubenswrapper[4948]: E0220 09:05:07.498056 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b940cc80e090937cbff77b5ed4366ab50ebcea5b7644547c7ab3fa22e20c4eeb\": container with ID starting with b940cc80e090937cbff77b5ed4366ab50ebcea5b7644547c7ab3fa22e20c4eeb not found: ID does not exist" containerID="b940cc80e090937cbff77b5ed4366ab50ebcea5b7644547c7ab3fa22e20c4eeb" Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.498125 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b940cc80e090937cbff77b5ed4366ab50ebcea5b7644547c7ab3fa22e20c4eeb"} err="failed to get container status \"b940cc80e090937cbff77b5ed4366ab50ebcea5b7644547c7ab3fa22e20c4eeb\": rpc error: code = NotFound desc = could not find container \"b940cc80e090937cbff77b5ed4366ab50ebcea5b7644547c7ab3fa22e20c4eeb\": container with ID starting with b940cc80e090937cbff77b5ed4366ab50ebcea5b7644547c7ab3fa22e20c4eeb not found: ID does not exist" Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.498166 4948 scope.go:117] "RemoveContainer" containerID="274f49b6cdbfb4b1df5dfe4a37883e06663d4ef43a70cc2e8f1bc96bbdbcfa76" Feb 20 09:05:07 crc kubenswrapper[4948]: E0220 09:05:07.498538 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"274f49b6cdbfb4b1df5dfe4a37883e06663d4ef43a70cc2e8f1bc96bbdbcfa76\": container with ID starting with 274f49b6cdbfb4b1df5dfe4a37883e06663d4ef43a70cc2e8f1bc96bbdbcfa76 not found: ID does not exist" containerID="274f49b6cdbfb4b1df5dfe4a37883e06663d4ef43a70cc2e8f1bc96bbdbcfa76" Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.498635 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"274f49b6cdbfb4b1df5dfe4a37883e06663d4ef43a70cc2e8f1bc96bbdbcfa76"} err="failed to get container status \"274f49b6cdbfb4b1df5dfe4a37883e06663d4ef43a70cc2e8f1bc96bbdbcfa76\": rpc error: code = NotFound desc = could not find container \"274f49b6cdbfb4b1df5dfe4a37883e06663d4ef43a70cc2e8f1bc96bbdbcfa76\": container with ID starting with 274f49b6cdbfb4b1df5dfe4a37883e06663d4ef43a70cc2e8f1bc96bbdbcfa76 not found: ID does not exist" Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.734826 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27f74a12-0843-4c61-a5c5-7f026224d86e" path="/var/lib/kubelet/pods/27f74a12-0843-4c61-a5c5-7f026224d86e/volumes" Feb 20 09:05:08 crc kubenswrapper[4948]: I0220 09:05:08.025054 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:05:08 crc kubenswrapper[4948]: I0220 09:05:08.025500 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:05:08 crc kubenswrapper[4948]: I0220 09:05:08.025564 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 09:05:08 crc kubenswrapper[4948]: I0220 09:05:08.026565 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2"} pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 09:05:08 crc kubenswrapper[4948]: I0220 09:05:08.026640 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" containerID="cri-o://6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" gracePeriod=600 Feb 20 09:05:08 crc kubenswrapper[4948]: E0220 09:05:08.169842 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:05:08 crc kubenswrapper[4948]: I0220 09:05:08.407250 4948 generic.go:334] "Generic (PLEG): container finished" podID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" exitCode=0 Feb 20 09:05:08 crc kubenswrapper[4948]: I0220 09:05:08.407316 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerDied","Data":"6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2"} Feb 20 09:05:08 crc kubenswrapper[4948]: I0220 09:05:08.407394 4948 scope.go:117] "RemoveContainer" containerID="b0e4da4a6a513a5c47c1f0dd87dcf0b56d3fede7e9008d839ac5623f61075ff8" Feb 20 09:05:08 crc kubenswrapper[4948]: I0220 09:05:08.408405 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:05:08 crc kubenswrapper[4948]: E0220 09:05:08.408881 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:05:11 crc kubenswrapper[4948]: I0220 09:05:11.498718 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-8sqsd/must-gather-xzw72"] Feb 20 09:05:11 crc kubenswrapper[4948]: I0220 09:05:11.499308 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-8sqsd/must-gather-xzw72" podUID="b25c36d5-5de9-4c98-acf4-495d492def68" containerName="copy" containerID="cri-o://ee095f93fada3e828226e407295d86f2e85f6d70e937ebcd69864d72a4cfb8b4" gracePeriod=2 Feb 20 09:05:11 crc kubenswrapper[4948]: I0220 09:05:11.520843 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-8sqsd/must-gather-xzw72"] Feb 20 09:05:11 crc kubenswrapper[4948]: I0220 09:05:11.945543 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-8sqsd_must-gather-xzw72_b25c36d5-5de9-4c98-acf4-495d492def68/copy/0.log" Feb 20 09:05:11 crc kubenswrapper[4948]: I0220 09:05:11.946344 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8sqsd/must-gather-xzw72" Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.024841 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsr8v\" (UniqueName: \"kubernetes.io/projected/b25c36d5-5de9-4c98-acf4-495d492def68-kube-api-access-vsr8v\") pod \"b25c36d5-5de9-4c98-acf4-495d492def68\" (UID: \"b25c36d5-5de9-4c98-acf4-495d492def68\") " Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.024929 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b25c36d5-5de9-4c98-acf4-495d492def68-must-gather-output\") pod \"b25c36d5-5de9-4c98-acf4-495d492def68\" (UID: \"b25c36d5-5de9-4c98-acf4-495d492def68\") " Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.030802 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b25c36d5-5de9-4c98-acf4-495d492def68-kube-api-access-vsr8v" (OuterVolumeSpecName: "kube-api-access-vsr8v") pod "b25c36d5-5de9-4c98-acf4-495d492def68" (UID: "b25c36d5-5de9-4c98-acf4-495d492def68"). InnerVolumeSpecName "kube-api-access-vsr8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.127529 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsr8v\" (UniqueName: \"kubernetes.io/projected/b25c36d5-5de9-4c98-acf4-495d492def68-kube-api-access-vsr8v\") on node \"crc\" DevicePath \"\"" Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.163937 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b25c36d5-5de9-4c98-acf4-495d492def68-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "b25c36d5-5de9-4c98-acf4-495d492def68" (UID: "b25c36d5-5de9-4c98-acf4-495d492def68"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.229669 4948 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b25c36d5-5de9-4c98-acf4-495d492def68-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.428316 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gs7f6" Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.428397 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gs7f6" Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.492333 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-8sqsd_must-gather-xzw72_b25c36d5-5de9-4c98-acf4-495d492def68/copy/0.log" Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.493161 4948 generic.go:334] "Generic (PLEG): container finished" podID="b25c36d5-5de9-4c98-acf4-495d492def68" containerID="ee095f93fada3e828226e407295d86f2e85f6d70e937ebcd69864d72a4cfb8b4" exitCode=143 Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.493261 4948 scope.go:117] "RemoveContainer" containerID="ee095f93fada3e828226e407295d86f2e85f6d70e937ebcd69864d72a4cfb8b4" Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.493276 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8sqsd/must-gather-xzw72" Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.530294 4948 scope.go:117] "RemoveContainer" containerID="4a8285325687a203e4a7cd75d1ec55e566e63219bb19f28c257e627288bb520a" Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.530403 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gs7f6" Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.599302 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gs7f6" Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.637938 4948 scope.go:117] "RemoveContainer" containerID="ee095f93fada3e828226e407295d86f2e85f6d70e937ebcd69864d72a4cfb8b4" Feb 20 09:05:12 crc kubenswrapper[4948]: E0220 09:05:12.638592 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee095f93fada3e828226e407295d86f2e85f6d70e937ebcd69864d72a4cfb8b4\": container with ID starting with ee095f93fada3e828226e407295d86f2e85f6d70e937ebcd69864d72a4cfb8b4 not found: ID does not exist" containerID="ee095f93fada3e828226e407295d86f2e85f6d70e937ebcd69864d72a4cfb8b4" Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.638635 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee095f93fada3e828226e407295d86f2e85f6d70e937ebcd69864d72a4cfb8b4"} err="failed to get container status \"ee095f93fada3e828226e407295d86f2e85f6d70e937ebcd69864d72a4cfb8b4\": rpc error: code = NotFound desc = could not find container \"ee095f93fada3e828226e407295d86f2e85f6d70e937ebcd69864d72a4cfb8b4\": container with ID starting with ee095f93fada3e828226e407295d86f2e85f6d70e937ebcd69864d72a4cfb8b4 not found: ID does not exist" Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.638660 4948 scope.go:117] "RemoveContainer" containerID="4a8285325687a203e4a7cd75d1ec55e566e63219bb19f28c257e627288bb520a" Feb 20 09:05:12 crc kubenswrapper[4948]: E0220 09:05:12.639163 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a8285325687a203e4a7cd75d1ec55e566e63219bb19f28c257e627288bb520a\": container with ID starting with 4a8285325687a203e4a7cd75d1ec55e566e63219bb19f28c257e627288bb520a not found: ID does not exist" containerID="4a8285325687a203e4a7cd75d1ec55e566e63219bb19f28c257e627288bb520a" Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.639205 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a8285325687a203e4a7cd75d1ec55e566e63219bb19f28c257e627288bb520a"} err="failed to get container status \"4a8285325687a203e4a7cd75d1ec55e566e63219bb19f28c257e627288bb520a\": rpc error: code = NotFound desc = could not find container \"4a8285325687a203e4a7cd75d1ec55e566e63219bb19f28c257e627288bb520a\": container with ID starting with 4a8285325687a203e4a7cd75d1ec55e566e63219bb19f28c257e627288bb520a not found: ID does not exist" Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.782284 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gs7f6"] Feb 20 09:05:13 crc kubenswrapper[4948]: I0220 09:05:13.734795 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b25c36d5-5de9-4c98-acf4-495d492def68" path="/var/lib/kubelet/pods/b25c36d5-5de9-4c98-acf4-495d492def68/volumes" Feb 20 09:05:14 crc kubenswrapper[4948]: I0220 09:05:14.523124 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gs7f6" podUID="fb1bc246-7f5a-4060-8bdb-6a3963ea6862" containerName="registry-server" containerID="cri-o://85b52de1041bebd75c7433b32671f81a45c491a3db70365771075aac31cce321" gracePeriod=2 Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.040628 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gs7f6" Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.086051 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-catalog-content\") pod \"fb1bc246-7f5a-4060-8bdb-6a3963ea6862\" (UID: \"fb1bc246-7f5a-4060-8bdb-6a3963ea6862\") " Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.086186 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-utilities\") pod \"fb1bc246-7f5a-4060-8bdb-6a3963ea6862\" (UID: \"fb1bc246-7f5a-4060-8bdb-6a3963ea6862\") " Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.086230 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmpwq\" (UniqueName: \"kubernetes.io/projected/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-kube-api-access-bmpwq\") pod \"fb1bc246-7f5a-4060-8bdb-6a3963ea6862\" (UID: \"fb1bc246-7f5a-4060-8bdb-6a3963ea6862\") " Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.088811 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-utilities" (OuterVolumeSpecName: "utilities") pod "fb1bc246-7f5a-4060-8bdb-6a3963ea6862" (UID: "fb1bc246-7f5a-4060-8bdb-6a3963ea6862"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.093184 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-kube-api-access-bmpwq" (OuterVolumeSpecName: "kube-api-access-bmpwq") pod "fb1bc246-7f5a-4060-8bdb-6a3963ea6862" (UID: "fb1bc246-7f5a-4060-8bdb-6a3963ea6862"). InnerVolumeSpecName "kube-api-access-bmpwq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.155924 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fb1bc246-7f5a-4060-8bdb-6a3963ea6862" (UID: "fb1bc246-7f5a-4060-8bdb-6a3963ea6862"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.188794 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.188868 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.188881 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmpwq\" (UniqueName: \"kubernetes.io/projected/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-kube-api-access-bmpwq\") on node \"crc\" DevicePath \"\"" Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.540246 4948 generic.go:334] "Generic (PLEG): container finished" podID="fb1bc246-7f5a-4060-8bdb-6a3963ea6862" containerID="85b52de1041bebd75c7433b32671f81a45c491a3db70365771075aac31cce321" exitCode=0 Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.540359 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gs7f6" event={"ID":"fb1bc246-7f5a-4060-8bdb-6a3963ea6862","Type":"ContainerDied","Data":"85b52de1041bebd75c7433b32671f81a45c491a3db70365771075aac31cce321"} Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.540412 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gs7f6" event={"ID":"fb1bc246-7f5a-4060-8bdb-6a3963ea6862","Type":"ContainerDied","Data":"a09d366d7a1bea20bacda46f1149340ebdb8bdbac35df3dd60174f0f350bbfa7"} Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.540452 4948 scope.go:117] "RemoveContainer" containerID="85b52de1041bebd75c7433b32671f81a45c491a3db70365771075aac31cce321" Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.540684 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gs7f6" Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.572814 4948 scope.go:117] "RemoveContainer" containerID="d5289ac930f2ff05d94bb857ba6dd563c33915f3a369ab91433c852e800b4d85" Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.598033 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gs7f6"] Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.607230 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gs7f6"] Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.614680 4948 scope.go:117] "RemoveContainer" containerID="ac1cf50592cbedded2cb785f164d800b487c31663eb504bce695abf3f1e85685" Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.665222 4948 scope.go:117] "RemoveContainer" containerID="85b52de1041bebd75c7433b32671f81a45c491a3db70365771075aac31cce321" Feb 20 09:05:15 crc kubenswrapper[4948]: E0220 09:05:15.665738 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85b52de1041bebd75c7433b32671f81a45c491a3db70365771075aac31cce321\": container with ID starting with 85b52de1041bebd75c7433b32671f81a45c491a3db70365771075aac31cce321 not found: ID does not exist" containerID="85b52de1041bebd75c7433b32671f81a45c491a3db70365771075aac31cce321" Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.665812 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85b52de1041bebd75c7433b32671f81a45c491a3db70365771075aac31cce321"} err="failed to get container status \"85b52de1041bebd75c7433b32671f81a45c491a3db70365771075aac31cce321\": rpc error: code = NotFound desc = could not find container \"85b52de1041bebd75c7433b32671f81a45c491a3db70365771075aac31cce321\": container with ID starting with 85b52de1041bebd75c7433b32671f81a45c491a3db70365771075aac31cce321 not found: ID does not exist" Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.665852 4948 scope.go:117] "RemoveContainer" containerID="d5289ac930f2ff05d94bb857ba6dd563c33915f3a369ab91433c852e800b4d85" Feb 20 09:05:15 crc kubenswrapper[4948]: E0220 09:05:15.666494 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5289ac930f2ff05d94bb857ba6dd563c33915f3a369ab91433c852e800b4d85\": container with ID starting with d5289ac930f2ff05d94bb857ba6dd563c33915f3a369ab91433c852e800b4d85 not found: ID does not exist" containerID="d5289ac930f2ff05d94bb857ba6dd563c33915f3a369ab91433c852e800b4d85" Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.666539 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5289ac930f2ff05d94bb857ba6dd563c33915f3a369ab91433c852e800b4d85"} err="failed to get container status \"d5289ac930f2ff05d94bb857ba6dd563c33915f3a369ab91433c852e800b4d85\": rpc error: code = NotFound desc = could not find container \"d5289ac930f2ff05d94bb857ba6dd563c33915f3a369ab91433c852e800b4d85\": container with ID starting with d5289ac930f2ff05d94bb857ba6dd563c33915f3a369ab91433c852e800b4d85 not found: ID does not exist" Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.666566 4948 scope.go:117] "RemoveContainer" containerID="ac1cf50592cbedded2cb785f164d800b487c31663eb504bce695abf3f1e85685" Feb 20 09:05:15 crc kubenswrapper[4948]: E0220 09:05:15.667140 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac1cf50592cbedded2cb785f164d800b487c31663eb504bce695abf3f1e85685\": container with ID starting with ac1cf50592cbedded2cb785f164d800b487c31663eb504bce695abf3f1e85685 not found: ID does not exist" containerID="ac1cf50592cbedded2cb785f164d800b487c31663eb504bce695abf3f1e85685" Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.667202 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac1cf50592cbedded2cb785f164d800b487c31663eb504bce695abf3f1e85685"} err="failed to get container status \"ac1cf50592cbedded2cb785f164d800b487c31663eb504bce695abf3f1e85685\": rpc error: code = NotFound desc = could not find container \"ac1cf50592cbedded2cb785f164d800b487c31663eb504bce695abf3f1e85685\": container with ID starting with ac1cf50592cbedded2cb785f164d800b487c31663eb504bce695abf3f1e85685 not found: ID does not exist" Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.745472 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb1bc246-7f5a-4060-8bdb-6a3963ea6862" path="/var/lib/kubelet/pods/fb1bc246-7f5a-4060-8bdb-6a3963ea6862/volumes" Feb 20 09:05:23 crc kubenswrapper[4948]: I0220 09:05:23.723147 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:05:23 crc kubenswrapper[4948]: E0220 09:05:23.724078 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:05:38 crc kubenswrapper[4948]: I0220 09:05:38.723125 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:05:38 crc kubenswrapper[4948]: E0220 09:05:38.724399 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:05:51 crc kubenswrapper[4948]: I0220 09:05:51.723355 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:05:51 crc kubenswrapper[4948]: E0220 09:05:51.724440 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:06:02 crc kubenswrapper[4948]: I0220 09:06:02.723368 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:06:02 crc kubenswrapper[4948]: E0220 09:06:02.724527 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:06:03 crc kubenswrapper[4948]: I0220 09:06:03.021094 4948 scope.go:117] "RemoveContainer" containerID="d2d5efbac326ed9a4b4067346df3dcafa9289ae45429a1d4be737ea540693745" Feb 20 09:06:17 crc kubenswrapper[4948]: I0220 09:06:17.722857 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:06:17 crc kubenswrapper[4948]: E0220 09:06:17.730222 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:06:29 crc kubenswrapper[4948]: I0220 09:06:29.722600 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:06:29 crc kubenswrapper[4948]: E0220 09:06:29.723427 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:06:41 crc kubenswrapper[4948]: I0220 09:06:41.734804 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:06:41 crc kubenswrapper[4948]: E0220 09:06:41.735721 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:06:56 crc kubenswrapper[4948]: I0220 09:06:56.722855 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:06:56 crc kubenswrapper[4948]: E0220 09:06:56.724942 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:07:08 crc kubenswrapper[4948]: I0220 09:07:08.722401 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:07:08 crc kubenswrapper[4948]: E0220 09:07:08.723392 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:07:21 crc kubenswrapper[4948]: I0220 09:07:21.740918 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:07:21 crc kubenswrapper[4948]: E0220 09:07:21.742387 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:07:33 crc kubenswrapper[4948]: I0220 09:07:33.722607 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:07:33 crc kubenswrapper[4948]: E0220 09:07:33.723716 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:07:44 crc kubenswrapper[4948]: I0220 09:07:44.723088 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:07:44 crc kubenswrapper[4948]: E0220 09:07:44.723954 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:07:56 crc kubenswrapper[4948]: I0220 09:07:56.722757 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:07:56 crc kubenswrapper[4948]: E0220 09:07:56.723684 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:08:07 crc kubenswrapper[4948]: I0220 09:08:07.722692 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:08:07 crc kubenswrapper[4948]: E0220 09:08:07.723663 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.782239 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ql8t9/must-gather-hvzrp"] Feb 20 09:08:10 crc kubenswrapper[4948]: E0220 09:08:10.784444 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27f74a12-0843-4c61-a5c5-7f026224d86e" containerName="registry-server" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.784571 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="27f74a12-0843-4c61-a5c5-7f026224d86e" containerName="registry-server" Feb 20 09:08:10 crc kubenswrapper[4948]: E0220 09:08:10.784677 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b25c36d5-5de9-4c98-acf4-495d492def68" containerName="copy" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.784768 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b25c36d5-5de9-4c98-acf4-495d492def68" containerName="copy" Feb 20 09:08:10 crc kubenswrapper[4948]: E0220 09:08:10.784891 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b25c36d5-5de9-4c98-acf4-495d492def68" containerName="gather" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.785002 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b25c36d5-5de9-4c98-acf4-495d492def68" containerName="gather" Feb 20 09:08:10 crc kubenswrapper[4948]: E0220 09:08:10.785115 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27f74a12-0843-4c61-a5c5-7f026224d86e" containerName="extract-utilities" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.785200 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="27f74a12-0843-4c61-a5c5-7f026224d86e" containerName="extract-utilities" Feb 20 09:08:10 crc kubenswrapper[4948]: E0220 09:08:10.785300 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27f74a12-0843-4c61-a5c5-7f026224d86e" containerName="extract-content" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.785387 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="27f74a12-0843-4c61-a5c5-7f026224d86e" containerName="extract-content" Feb 20 09:08:10 crc kubenswrapper[4948]: E0220 09:08:10.785488 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb1bc246-7f5a-4060-8bdb-6a3963ea6862" containerName="extract-content" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.785588 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb1bc246-7f5a-4060-8bdb-6a3963ea6862" containerName="extract-content" Feb 20 09:08:10 crc kubenswrapper[4948]: E0220 09:08:10.785678 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb1bc246-7f5a-4060-8bdb-6a3963ea6862" containerName="extract-utilities" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.785763 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb1bc246-7f5a-4060-8bdb-6a3963ea6862" containerName="extract-utilities" Feb 20 09:08:10 crc kubenswrapper[4948]: E0220 09:08:10.785845 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb1bc246-7f5a-4060-8bdb-6a3963ea6862" containerName="registry-server" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.785929 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb1bc246-7f5a-4060-8bdb-6a3963ea6862" containerName="registry-server" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.786361 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="b25c36d5-5de9-4c98-acf4-495d492def68" containerName="gather" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.786487 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb1bc246-7f5a-4060-8bdb-6a3963ea6862" containerName="registry-server" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.786633 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="27f74a12-0843-4c61-a5c5-7f026224d86e" containerName="registry-server" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.786738 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="b25c36d5-5de9-4c98-acf4-495d492def68" containerName="copy" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.788128 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ql8t9/must-gather-hvzrp" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.790162 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-ql8t9/must-gather-hvzrp"] Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.790887 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-ql8t9"/"kube-root-ca.crt" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.790980 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-ql8t9"/"openshift-service-ca.crt" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.920335 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2bcm\" (UniqueName: \"kubernetes.io/projected/d81b349b-2494-4abc-b349-44fecd284709-kube-api-access-v2bcm\") pod \"must-gather-hvzrp\" (UID: \"d81b349b-2494-4abc-b349-44fecd284709\") " pod="openshift-must-gather-ql8t9/must-gather-hvzrp" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.920379 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d81b349b-2494-4abc-b349-44fecd284709-must-gather-output\") pod \"must-gather-hvzrp\" (UID: \"d81b349b-2494-4abc-b349-44fecd284709\") " pod="openshift-must-gather-ql8t9/must-gather-hvzrp" Feb 20 09:08:11 crc kubenswrapper[4948]: I0220 09:08:11.022503 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2bcm\" (UniqueName: \"kubernetes.io/projected/d81b349b-2494-4abc-b349-44fecd284709-kube-api-access-v2bcm\") pod \"must-gather-hvzrp\" (UID: \"d81b349b-2494-4abc-b349-44fecd284709\") " pod="openshift-must-gather-ql8t9/must-gather-hvzrp" Feb 20 09:08:11 crc kubenswrapper[4948]: I0220 09:08:11.022571 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d81b349b-2494-4abc-b349-44fecd284709-must-gather-output\") pod \"must-gather-hvzrp\" (UID: \"d81b349b-2494-4abc-b349-44fecd284709\") " pod="openshift-must-gather-ql8t9/must-gather-hvzrp" Feb 20 09:08:11 crc kubenswrapper[4948]: I0220 09:08:11.023303 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d81b349b-2494-4abc-b349-44fecd284709-must-gather-output\") pod \"must-gather-hvzrp\" (UID: \"d81b349b-2494-4abc-b349-44fecd284709\") " pod="openshift-must-gather-ql8t9/must-gather-hvzrp" Feb 20 09:08:11 crc kubenswrapper[4948]: I0220 09:08:11.049897 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2bcm\" (UniqueName: \"kubernetes.io/projected/d81b349b-2494-4abc-b349-44fecd284709-kube-api-access-v2bcm\") pod \"must-gather-hvzrp\" (UID: \"d81b349b-2494-4abc-b349-44fecd284709\") " pod="openshift-must-gather-ql8t9/must-gather-hvzrp" Feb 20 09:08:11 crc kubenswrapper[4948]: I0220 09:08:11.110308 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ql8t9/must-gather-hvzrp" Feb 20 09:08:11 crc kubenswrapper[4948]: I0220 09:08:11.549867 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-ql8t9/must-gather-hvzrp"] Feb 20 09:08:11 crc kubenswrapper[4948]: W0220 09:08:11.550241 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd81b349b_2494_4abc_b349_44fecd284709.slice/crio-e4d5c7dd67fd11d407c5c350d2cbad2c4a79d8cf80019f91b10ec82e4e0a0226 WatchSource:0}: Error finding container e4d5c7dd67fd11d407c5c350d2cbad2c4a79d8cf80019f91b10ec82e4e0a0226: Status 404 returned error can't find the container with id e4d5c7dd67fd11d407c5c350d2cbad2c4a79d8cf80019f91b10ec82e4e0a0226 Feb 20 09:08:11 crc kubenswrapper[4948]: I0220 09:08:11.587683 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ql8t9/must-gather-hvzrp" event={"ID":"d81b349b-2494-4abc-b349-44fecd284709","Type":"ContainerStarted","Data":"e4d5c7dd67fd11d407c5c350d2cbad2c4a79d8cf80019f91b10ec82e4e0a0226"} Feb 20 09:08:12 crc kubenswrapper[4948]: I0220 09:08:12.599244 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ql8t9/must-gather-hvzrp" event={"ID":"d81b349b-2494-4abc-b349-44fecd284709","Type":"ContainerStarted","Data":"d49cb379a847549834cd028d6c373359633cf1e9d6c301526cf59089fd2bfebf"} Feb 20 09:08:12 crc kubenswrapper[4948]: I0220 09:08:12.599570 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ql8t9/must-gather-hvzrp" event={"ID":"d81b349b-2494-4abc-b349-44fecd284709","Type":"ContainerStarted","Data":"0b5814bb586cbca972e209566bf544539a27cba45472fac01d96fe47fe540cff"} Feb 20 09:08:12 crc kubenswrapper[4948]: I0220 09:08:12.622631 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-ql8t9/must-gather-hvzrp" podStartSLOduration=2.622603205 podStartE2EDuration="2.622603205s" podCreationTimestamp="2026-02-20 09:08:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 09:08:12.616960336 +0000 UTC m=+3741.591455146" watchObservedRunningTime="2026-02-20 09:08:12.622603205 +0000 UTC m=+3741.597098045" Feb 20 09:08:15 crc kubenswrapper[4948]: E0220 09:08:15.150365 4948 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.113:47142->38.102.83.113:38489: write tcp 38.102.83.113:47142->38.102.83.113:38489: write: broken pipe Feb 20 09:08:15 crc kubenswrapper[4948]: I0220 09:08:15.825765 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ql8t9/crc-debug-rdtwt"] Feb 20 09:08:15 crc kubenswrapper[4948]: I0220 09:08:15.826840 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ql8t9/crc-debug-rdtwt" Feb 20 09:08:15 crc kubenswrapper[4948]: I0220 09:08:15.828777 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-ql8t9"/"default-dockercfg-n7kqw" Feb 20 09:08:15 crc kubenswrapper[4948]: I0220 09:08:15.924603 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsf25\" (UniqueName: \"kubernetes.io/projected/25487e7a-c3d7-47c0-9ef0-1bd55109725a-kube-api-access-nsf25\") pod \"crc-debug-rdtwt\" (UID: \"25487e7a-c3d7-47c0-9ef0-1bd55109725a\") " pod="openshift-must-gather-ql8t9/crc-debug-rdtwt" Feb 20 09:08:15 crc kubenswrapper[4948]: I0220 09:08:15.924885 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/25487e7a-c3d7-47c0-9ef0-1bd55109725a-host\") pod \"crc-debug-rdtwt\" (UID: \"25487e7a-c3d7-47c0-9ef0-1bd55109725a\") " pod="openshift-must-gather-ql8t9/crc-debug-rdtwt" Feb 20 09:08:16 crc kubenswrapper[4948]: I0220 09:08:16.027210 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsf25\" (UniqueName: \"kubernetes.io/projected/25487e7a-c3d7-47c0-9ef0-1bd55109725a-kube-api-access-nsf25\") pod \"crc-debug-rdtwt\" (UID: \"25487e7a-c3d7-47c0-9ef0-1bd55109725a\") " pod="openshift-must-gather-ql8t9/crc-debug-rdtwt" Feb 20 09:08:16 crc kubenswrapper[4948]: I0220 09:08:16.027841 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/25487e7a-c3d7-47c0-9ef0-1bd55109725a-host\") pod \"crc-debug-rdtwt\" (UID: \"25487e7a-c3d7-47c0-9ef0-1bd55109725a\") " pod="openshift-must-gather-ql8t9/crc-debug-rdtwt" Feb 20 09:08:16 crc kubenswrapper[4948]: I0220 09:08:16.027954 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/25487e7a-c3d7-47c0-9ef0-1bd55109725a-host\") pod \"crc-debug-rdtwt\" (UID: \"25487e7a-c3d7-47c0-9ef0-1bd55109725a\") " pod="openshift-must-gather-ql8t9/crc-debug-rdtwt" Feb 20 09:08:16 crc kubenswrapper[4948]: I0220 09:08:16.046302 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsf25\" (UniqueName: \"kubernetes.io/projected/25487e7a-c3d7-47c0-9ef0-1bd55109725a-kube-api-access-nsf25\") pod \"crc-debug-rdtwt\" (UID: \"25487e7a-c3d7-47c0-9ef0-1bd55109725a\") " pod="openshift-must-gather-ql8t9/crc-debug-rdtwt" Feb 20 09:08:16 crc kubenswrapper[4948]: I0220 09:08:16.143498 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ql8t9/crc-debug-rdtwt" Feb 20 09:08:16 crc kubenswrapper[4948]: I0220 09:08:16.637731 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ql8t9/crc-debug-rdtwt" event={"ID":"25487e7a-c3d7-47c0-9ef0-1bd55109725a","Type":"ContainerStarted","Data":"7079ff1f2f1f98bbc8f43f1d268a42718af20f413eed66d493a146a8c4b11eb9"} Feb 20 09:08:16 crc kubenswrapper[4948]: I0220 09:08:16.639155 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ql8t9/crc-debug-rdtwt" event={"ID":"25487e7a-c3d7-47c0-9ef0-1bd55109725a","Type":"ContainerStarted","Data":"96045d0fbf4acb31ef0e8ddf75f287243e16367e34ffe86e106174591cc4fc32"} Feb 20 09:08:20 crc kubenswrapper[4948]: I0220 09:08:20.723202 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:08:20 crc kubenswrapper[4948]: E0220 09:08:20.723994 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:08:34 crc kubenswrapper[4948]: I0220 09:08:34.723818 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:08:34 crc kubenswrapper[4948]: E0220 09:08:34.724599 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:08:48 crc kubenswrapper[4948]: I0220 09:08:48.723798 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:08:48 crc kubenswrapper[4948]: E0220 09:08:48.725032 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:08:49 crc kubenswrapper[4948]: I0220 09:08:49.956827 4948 generic.go:334] "Generic (PLEG): container finished" podID="25487e7a-c3d7-47c0-9ef0-1bd55109725a" containerID="7079ff1f2f1f98bbc8f43f1d268a42718af20f413eed66d493a146a8c4b11eb9" exitCode=0 Feb 20 09:08:49 crc kubenswrapper[4948]: I0220 09:08:49.957230 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ql8t9/crc-debug-rdtwt" event={"ID":"25487e7a-c3d7-47c0-9ef0-1bd55109725a","Type":"ContainerDied","Data":"7079ff1f2f1f98bbc8f43f1d268a42718af20f413eed66d493a146a8c4b11eb9"} Feb 20 09:08:51 crc kubenswrapper[4948]: I0220 09:08:51.178300 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ql8t9/crc-debug-rdtwt" Feb 20 09:08:51 crc kubenswrapper[4948]: I0220 09:08:51.218382 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ql8t9/crc-debug-rdtwt"] Feb 20 09:08:51 crc kubenswrapper[4948]: I0220 09:08:51.230525 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ql8t9/crc-debug-rdtwt"] Feb 20 09:08:51 crc kubenswrapper[4948]: I0220 09:08:51.303967 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nsf25\" (UniqueName: \"kubernetes.io/projected/25487e7a-c3d7-47c0-9ef0-1bd55109725a-kube-api-access-nsf25\") pod \"25487e7a-c3d7-47c0-9ef0-1bd55109725a\" (UID: \"25487e7a-c3d7-47c0-9ef0-1bd55109725a\") " Feb 20 09:08:51 crc kubenswrapper[4948]: I0220 09:08:51.304178 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/25487e7a-c3d7-47c0-9ef0-1bd55109725a-host\") pod \"25487e7a-c3d7-47c0-9ef0-1bd55109725a\" (UID: \"25487e7a-c3d7-47c0-9ef0-1bd55109725a\") " Feb 20 09:08:51 crc kubenswrapper[4948]: I0220 09:08:51.304654 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/25487e7a-c3d7-47c0-9ef0-1bd55109725a-host" (OuterVolumeSpecName: "host") pod "25487e7a-c3d7-47c0-9ef0-1bd55109725a" (UID: "25487e7a-c3d7-47c0-9ef0-1bd55109725a"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 09:08:51 crc kubenswrapper[4948]: I0220 09:08:51.309843 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25487e7a-c3d7-47c0-9ef0-1bd55109725a-kube-api-access-nsf25" (OuterVolumeSpecName: "kube-api-access-nsf25") pod "25487e7a-c3d7-47c0-9ef0-1bd55109725a" (UID: "25487e7a-c3d7-47c0-9ef0-1bd55109725a"). InnerVolumeSpecName "kube-api-access-nsf25". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:08:51 crc kubenswrapper[4948]: I0220 09:08:51.405908 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nsf25\" (UniqueName: \"kubernetes.io/projected/25487e7a-c3d7-47c0-9ef0-1bd55109725a-kube-api-access-nsf25\") on node \"crc\" DevicePath \"\"" Feb 20 09:08:51 crc kubenswrapper[4948]: I0220 09:08:51.405940 4948 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/25487e7a-c3d7-47c0-9ef0-1bd55109725a-host\") on node \"crc\" DevicePath \"\"" Feb 20 09:08:51 crc kubenswrapper[4948]: I0220 09:08:51.739041 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25487e7a-c3d7-47c0-9ef0-1bd55109725a" path="/var/lib/kubelet/pods/25487e7a-c3d7-47c0-9ef0-1bd55109725a/volumes" Feb 20 09:08:51 crc kubenswrapper[4948]: I0220 09:08:51.971819 4948 scope.go:117] "RemoveContainer" containerID="7079ff1f2f1f98bbc8f43f1d268a42718af20f413eed66d493a146a8c4b11eb9" Feb 20 09:08:51 crc kubenswrapper[4948]: I0220 09:08:51.971853 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ql8t9/crc-debug-rdtwt" Feb 20 09:08:52 crc kubenswrapper[4948]: I0220 09:08:52.393958 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ql8t9/crc-debug-njlvs"] Feb 20 09:08:52 crc kubenswrapper[4948]: E0220 09:08:52.394305 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25487e7a-c3d7-47c0-9ef0-1bd55109725a" containerName="container-00" Feb 20 09:08:52 crc kubenswrapper[4948]: I0220 09:08:52.394317 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="25487e7a-c3d7-47c0-9ef0-1bd55109725a" containerName="container-00" Feb 20 09:08:52 crc kubenswrapper[4948]: I0220 09:08:52.394492 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="25487e7a-c3d7-47c0-9ef0-1bd55109725a" containerName="container-00" Feb 20 09:08:52 crc kubenswrapper[4948]: I0220 09:08:52.395098 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ql8t9/crc-debug-njlvs" Feb 20 09:08:52 crc kubenswrapper[4948]: I0220 09:08:52.398115 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-ql8t9"/"default-dockercfg-n7kqw" Feb 20 09:08:52 crc kubenswrapper[4948]: I0220 09:08:52.525684 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/946347cb-7ccd-45e9-866c-2f9b2d6e10cb-host\") pod \"crc-debug-njlvs\" (UID: \"946347cb-7ccd-45e9-866c-2f9b2d6e10cb\") " pod="openshift-must-gather-ql8t9/crc-debug-njlvs" Feb 20 09:08:52 crc kubenswrapper[4948]: I0220 09:08:52.525984 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjmgr\" (UniqueName: \"kubernetes.io/projected/946347cb-7ccd-45e9-866c-2f9b2d6e10cb-kube-api-access-cjmgr\") pod \"crc-debug-njlvs\" (UID: \"946347cb-7ccd-45e9-866c-2f9b2d6e10cb\") " pod="openshift-must-gather-ql8t9/crc-debug-njlvs" Feb 20 09:08:52 crc kubenswrapper[4948]: I0220 09:08:52.627553 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/946347cb-7ccd-45e9-866c-2f9b2d6e10cb-host\") pod \"crc-debug-njlvs\" (UID: \"946347cb-7ccd-45e9-866c-2f9b2d6e10cb\") " pod="openshift-must-gather-ql8t9/crc-debug-njlvs" Feb 20 09:08:52 crc kubenswrapper[4948]: I0220 09:08:52.627670 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjmgr\" (UniqueName: \"kubernetes.io/projected/946347cb-7ccd-45e9-866c-2f9b2d6e10cb-kube-api-access-cjmgr\") pod \"crc-debug-njlvs\" (UID: \"946347cb-7ccd-45e9-866c-2f9b2d6e10cb\") " pod="openshift-must-gather-ql8t9/crc-debug-njlvs" Feb 20 09:08:52 crc kubenswrapper[4948]: I0220 09:08:52.627769 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/946347cb-7ccd-45e9-866c-2f9b2d6e10cb-host\") pod \"crc-debug-njlvs\" (UID: \"946347cb-7ccd-45e9-866c-2f9b2d6e10cb\") " pod="openshift-must-gather-ql8t9/crc-debug-njlvs" Feb 20 09:08:52 crc kubenswrapper[4948]: I0220 09:08:52.648154 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjmgr\" (UniqueName: \"kubernetes.io/projected/946347cb-7ccd-45e9-866c-2f9b2d6e10cb-kube-api-access-cjmgr\") pod \"crc-debug-njlvs\" (UID: \"946347cb-7ccd-45e9-866c-2f9b2d6e10cb\") " pod="openshift-must-gather-ql8t9/crc-debug-njlvs" Feb 20 09:08:52 crc kubenswrapper[4948]: I0220 09:08:52.726609 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ql8t9/crc-debug-njlvs" Feb 20 09:08:52 crc kubenswrapper[4948]: W0220 09:08:52.789023 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod946347cb_7ccd_45e9_866c_2f9b2d6e10cb.slice/crio-dc66a6b6c2a4cedaf7ccafd526de87fb780940e37aeeeb99e15e84f58fc897f3 WatchSource:0}: Error finding container dc66a6b6c2a4cedaf7ccafd526de87fb780940e37aeeeb99e15e84f58fc897f3: Status 404 returned error can't find the container with id dc66a6b6c2a4cedaf7ccafd526de87fb780940e37aeeeb99e15e84f58fc897f3 Feb 20 09:08:52 crc kubenswrapper[4948]: I0220 09:08:52.981573 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ql8t9/crc-debug-njlvs" event={"ID":"946347cb-7ccd-45e9-866c-2f9b2d6e10cb","Type":"ContainerStarted","Data":"dc66a6b6c2a4cedaf7ccafd526de87fb780940e37aeeeb99e15e84f58fc897f3"} Feb 20 09:08:53 crc kubenswrapper[4948]: I0220 09:08:53.992356 4948 generic.go:334] "Generic (PLEG): container finished" podID="946347cb-7ccd-45e9-866c-2f9b2d6e10cb" containerID="01b0492d26413e240c32017d4eb44b4ec84b1e4c3a3eedc722410b0fbfdd91a9" exitCode=0 Feb 20 09:08:53 crc kubenswrapper[4948]: I0220 09:08:53.992403 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ql8t9/crc-debug-njlvs" event={"ID":"946347cb-7ccd-45e9-866c-2f9b2d6e10cb","Type":"ContainerDied","Data":"01b0492d26413e240c32017d4eb44b4ec84b1e4c3a3eedc722410b0fbfdd91a9"} Feb 20 09:08:54 crc kubenswrapper[4948]: I0220 09:08:54.428870 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ql8t9/crc-debug-njlvs"] Feb 20 09:08:54 crc kubenswrapper[4948]: I0220 09:08:54.440280 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ql8t9/crc-debug-njlvs"] Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.088411 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ql8t9/crc-debug-njlvs" Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.171699 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjmgr\" (UniqueName: \"kubernetes.io/projected/946347cb-7ccd-45e9-866c-2f9b2d6e10cb-kube-api-access-cjmgr\") pod \"946347cb-7ccd-45e9-866c-2f9b2d6e10cb\" (UID: \"946347cb-7ccd-45e9-866c-2f9b2d6e10cb\") " Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.171904 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/946347cb-7ccd-45e9-866c-2f9b2d6e10cb-host\") pod \"946347cb-7ccd-45e9-866c-2f9b2d6e10cb\" (UID: \"946347cb-7ccd-45e9-866c-2f9b2d6e10cb\") " Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.172049 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/946347cb-7ccd-45e9-866c-2f9b2d6e10cb-host" (OuterVolumeSpecName: "host") pod "946347cb-7ccd-45e9-866c-2f9b2d6e10cb" (UID: "946347cb-7ccd-45e9-866c-2f9b2d6e10cb"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.172354 4948 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/946347cb-7ccd-45e9-866c-2f9b2d6e10cb-host\") on node \"crc\" DevicePath \"\"" Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.184722 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/946347cb-7ccd-45e9-866c-2f9b2d6e10cb-kube-api-access-cjmgr" (OuterVolumeSpecName: "kube-api-access-cjmgr") pod "946347cb-7ccd-45e9-866c-2f9b2d6e10cb" (UID: "946347cb-7ccd-45e9-866c-2f9b2d6e10cb"). InnerVolumeSpecName "kube-api-access-cjmgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.274499 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjmgr\" (UniqueName: \"kubernetes.io/projected/946347cb-7ccd-45e9-866c-2f9b2d6e10cb-kube-api-access-cjmgr\") on node \"crc\" DevicePath \"\"" Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.657787 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ql8t9/crc-debug-5gtxj"] Feb 20 09:08:55 crc kubenswrapper[4948]: E0220 09:08:55.658171 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="946347cb-7ccd-45e9-866c-2f9b2d6e10cb" containerName="container-00" Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.658189 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="946347cb-7ccd-45e9-866c-2f9b2d6e10cb" containerName="container-00" Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.658367 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="946347cb-7ccd-45e9-866c-2f9b2d6e10cb" containerName="container-00" Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.658950 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ql8t9/crc-debug-5gtxj" Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.732094 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="946347cb-7ccd-45e9-866c-2f9b2d6e10cb" path="/var/lib/kubelet/pods/946347cb-7ccd-45e9-866c-2f9b2d6e10cb/volumes" Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.785412 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t82dt\" (UniqueName: \"kubernetes.io/projected/0cfde212-7150-4b57-b3dd-41715c736426-kube-api-access-t82dt\") pod \"crc-debug-5gtxj\" (UID: \"0cfde212-7150-4b57-b3dd-41715c736426\") " pod="openshift-must-gather-ql8t9/crc-debug-5gtxj" Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.785725 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0cfde212-7150-4b57-b3dd-41715c736426-host\") pod \"crc-debug-5gtxj\" (UID: \"0cfde212-7150-4b57-b3dd-41715c736426\") " pod="openshift-must-gather-ql8t9/crc-debug-5gtxj" Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.887907 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t82dt\" (UniqueName: \"kubernetes.io/projected/0cfde212-7150-4b57-b3dd-41715c736426-kube-api-access-t82dt\") pod \"crc-debug-5gtxj\" (UID: \"0cfde212-7150-4b57-b3dd-41715c736426\") " pod="openshift-must-gather-ql8t9/crc-debug-5gtxj" Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.888612 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0cfde212-7150-4b57-b3dd-41715c736426-host\") pod \"crc-debug-5gtxj\" (UID: \"0cfde212-7150-4b57-b3dd-41715c736426\") " pod="openshift-must-gather-ql8t9/crc-debug-5gtxj" Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.888443 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0cfde212-7150-4b57-b3dd-41715c736426-host\") pod \"crc-debug-5gtxj\" (UID: \"0cfde212-7150-4b57-b3dd-41715c736426\") " pod="openshift-must-gather-ql8t9/crc-debug-5gtxj" Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.905889 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t82dt\" (UniqueName: \"kubernetes.io/projected/0cfde212-7150-4b57-b3dd-41715c736426-kube-api-access-t82dt\") pod \"crc-debug-5gtxj\" (UID: \"0cfde212-7150-4b57-b3dd-41715c736426\") " pod="openshift-must-gather-ql8t9/crc-debug-5gtxj" Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.974357 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ql8t9/crc-debug-5gtxj" Feb 20 09:08:56 crc kubenswrapper[4948]: I0220 09:08:56.030173 4948 scope.go:117] "RemoveContainer" containerID="01b0492d26413e240c32017d4eb44b4ec84b1e4c3a3eedc722410b0fbfdd91a9" Feb 20 09:08:56 crc kubenswrapper[4948]: I0220 09:08:56.030308 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ql8t9/crc-debug-njlvs" Feb 20 09:08:56 crc kubenswrapper[4948]: I0220 09:08:56.042116 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ql8t9/crc-debug-5gtxj" event={"ID":"0cfde212-7150-4b57-b3dd-41715c736426","Type":"ContainerStarted","Data":"6f9248e9b3243b7c97c4714220f5bdb0e400fa6c3ef7be9e1e20648a82dc1da6"} Feb 20 09:08:57 crc kubenswrapper[4948]: I0220 09:08:57.054345 4948 generic.go:334] "Generic (PLEG): container finished" podID="0cfde212-7150-4b57-b3dd-41715c736426" containerID="cf4963aa778b92fd84cff1d5614e4d02e1a3608c465ada5c0b7912612f7613de" exitCode=0 Feb 20 09:08:57 crc kubenswrapper[4948]: I0220 09:08:57.054418 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ql8t9/crc-debug-5gtxj" event={"ID":"0cfde212-7150-4b57-b3dd-41715c736426","Type":"ContainerDied","Data":"cf4963aa778b92fd84cff1d5614e4d02e1a3608c465ada5c0b7912612f7613de"} Feb 20 09:08:57 crc kubenswrapper[4948]: I0220 09:08:57.107942 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ql8t9/crc-debug-5gtxj"] Feb 20 09:08:57 crc kubenswrapper[4948]: I0220 09:08:57.117287 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ql8t9/crc-debug-5gtxj"] Feb 20 09:08:58 crc kubenswrapper[4948]: I0220 09:08:58.192697 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ql8t9/crc-debug-5gtxj" Feb 20 09:08:58 crc kubenswrapper[4948]: I0220 09:08:58.339256 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t82dt\" (UniqueName: \"kubernetes.io/projected/0cfde212-7150-4b57-b3dd-41715c736426-kube-api-access-t82dt\") pod \"0cfde212-7150-4b57-b3dd-41715c736426\" (UID: \"0cfde212-7150-4b57-b3dd-41715c736426\") " Feb 20 09:08:58 crc kubenswrapper[4948]: I0220 09:08:58.339430 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0cfde212-7150-4b57-b3dd-41715c736426-host\") pod \"0cfde212-7150-4b57-b3dd-41715c736426\" (UID: \"0cfde212-7150-4b57-b3dd-41715c736426\") " Feb 20 09:08:58 crc kubenswrapper[4948]: I0220 09:08:58.339563 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cfde212-7150-4b57-b3dd-41715c736426-host" (OuterVolumeSpecName: "host") pod "0cfde212-7150-4b57-b3dd-41715c736426" (UID: "0cfde212-7150-4b57-b3dd-41715c736426"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 09:08:58 crc kubenswrapper[4948]: I0220 09:08:58.339871 4948 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0cfde212-7150-4b57-b3dd-41715c736426-host\") on node \"crc\" DevicePath \"\"" Feb 20 09:08:58 crc kubenswrapper[4948]: I0220 09:08:58.344231 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cfde212-7150-4b57-b3dd-41715c736426-kube-api-access-t82dt" (OuterVolumeSpecName: "kube-api-access-t82dt") pod "0cfde212-7150-4b57-b3dd-41715c736426" (UID: "0cfde212-7150-4b57-b3dd-41715c736426"). InnerVolumeSpecName "kube-api-access-t82dt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:08:58 crc kubenswrapper[4948]: I0220 09:08:58.441401 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t82dt\" (UniqueName: \"kubernetes.io/projected/0cfde212-7150-4b57-b3dd-41715c736426-kube-api-access-t82dt\") on node \"crc\" DevicePath \"\"" Feb 20 09:08:59 crc kubenswrapper[4948]: I0220 09:08:59.078070 4948 scope.go:117] "RemoveContainer" containerID="cf4963aa778b92fd84cff1d5614e4d02e1a3608c465ada5c0b7912612f7613de" Feb 20 09:08:59 crc kubenswrapper[4948]: I0220 09:08:59.078351 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ql8t9/crc-debug-5gtxj" Feb 20 09:08:59 crc kubenswrapper[4948]: I0220 09:08:59.722235 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:08:59 crc kubenswrapper[4948]: E0220 09:08:59.722606 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:08:59 crc kubenswrapper[4948]: I0220 09:08:59.741744 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cfde212-7150-4b57-b3dd-41715c736426" path="/var/lib/kubelet/pods/0cfde212-7150-4b57-b3dd-41715c736426/volumes" Feb 20 09:09:11 crc kubenswrapper[4948]: I0220 09:09:11.728218 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:09:11 crc kubenswrapper[4948]: E0220 09:09:11.728940 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:09:24 crc kubenswrapper[4948]: I0220 09:09:24.723393 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:09:24 crc kubenswrapper[4948]: E0220 09:09:24.724370 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:09:30 crc kubenswrapper[4948]: I0220 09:09:30.743588 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-686d666564-8tnjb_51f6301c-57be-4a07-91df-b1a9f90e3f40/barbican-api/0.log" Feb 20 09:09:30 crc kubenswrapper[4948]: I0220 09:09:30.880665 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-686d666564-8tnjb_51f6301c-57be-4a07-91df-b1a9f90e3f40/barbican-api-log/0.log" Feb 20 09:09:30 crc kubenswrapper[4948]: I0220 09:09:30.939532 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6ffd468-xgk27_292d8e1e-de3b-4259-afef-c19e2c21187c/barbican-keystone-listener/0.log" Feb 20 09:09:31 crc kubenswrapper[4948]: I0220 09:09:31.026226 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6ffd468-xgk27_292d8e1e-de3b-4259-afef-c19e2c21187c/barbican-keystone-listener-log/0.log" Feb 20 09:09:31 crc kubenswrapper[4948]: I0220 09:09:31.122477 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-76d486c65f-w9lc6_4b356058-d957-4fa5-a106-3fe5052c8539/barbican-worker/0.log" Feb 20 09:09:31 crc kubenswrapper[4948]: I0220 09:09:31.187164 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-76d486c65f-w9lc6_4b356058-d957-4fa5-a106-3fe5052c8539/barbican-worker-log/0.log" Feb 20 09:09:31 crc kubenswrapper[4948]: I0220 09:09:31.314606 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-4265g_e9bed94f-dc85-433b-b2db-8da400959f54/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:09:31 crc kubenswrapper[4948]: I0220 09:09:31.416195 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_58f563bb-9716-4abd-b187-3f01036d4f31/ceilometer-central-agent/0.log" Feb 20 09:09:31 crc kubenswrapper[4948]: I0220 09:09:31.450624 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_58f563bb-9716-4abd-b187-3f01036d4f31/ceilometer-notification-agent/0.log" Feb 20 09:09:31 crc kubenswrapper[4948]: I0220 09:09:31.484522 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_58f563bb-9716-4abd-b187-3f01036d4f31/proxy-httpd/0.log" Feb 20 09:09:31 crc kubenswrapper[4948]: I0220 09:09:31.578693 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_58f563bb-9716-4abd-b187-3f01036d4f31/sg-core/0.log" Feb 20 09:09:31 crc kubenswrapper[4948]: I0220 09:09:31.677176 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_e5967cab-cc29-4c87-b873-0f5d20203d4f/cinder-api-log/0.log" Feb 20 09:09:31 crc kubenswrapper[4948]: I0220 09:09:31.682705 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_e5967cab-cc29-4c87-b873-0f5d20203d4f/cinder-api/0.log" Feb 20 09:09:31 crc kubenswrapper[4948]: I0220 09:09:31.890280 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2f8a0f7b-ee05-4e93-a591-7486fecf8cd0/cinder-scheduler/0.log" Feb 20 09:09:31 crc kubenswrapper[4948]: I0220 09:09:31.893314 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2f8a0f7b-ee05-4e93-a591-7486fecf8cd0/probe/0.log" Feb 20 09:09:32 crc kubenswrapper[4948]: I0220 09:09:32.095689 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx_bf3501bf-8f1e-4529-91a8-6be83eda4158/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:09:32 crc kubenswrapper[4948]: I0220 09:09:32.098565 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6_d38195a4-0c7b-427e-b5e2-923b24c10674/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:09:32 crc kubenswrapper[4948]: I0220 09:09:32.419807 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-fbjpk_b05f92c6-60c9-41be-bf6c-6288cee04659/init/0.log" Feb 20 09:09:32 crc kubenswrapper[4948]: I0220 09:09:32.614023 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-fbjpk_b05f92c6-60c9-41be-bf6c-6288cee04659/init/0.log" Feb 20 09:09:32 crc kubenswrapper[4948]: I0220 09:09:32.699614 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-mhkll_5864a68d-650e-4bcf-b705-619c0f27445b/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:09:32 crc kubenswrapper[4948]: I0220 09:09:32.702710 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-fbjpk_b05f92c6-60c9-41be-bf6c-6288cee04659/dnsmasq-dns/0.log" Feb 20 09:09:32 crc kubenswrapper[4948]: I0220 09:09:32.920952 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a/glance-log/0.log" Feb 20 09:09:32 crc kubenswrapper[4948]: I0220 09:09:32.970139 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a/glance-httpd/0.log" Feb 20 09:09:33 crc kubenswrapper[4948]: I0220 09:09:33.059730 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_4fcb965e-4d04-4863-a966-39a83f458fa6/glance-httpd/0.log" Feb 20 09:09:33 crc kubenswrapper[4948]: I0220 09:09:33.145268 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_4fcb965e-4d04-4863-a966-39a83f458fa6/glance-log/0.log" Feb 20 09:09:33 crc kubenswrapper[4948]: I0220 09:09:33.242401 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-685c4db87b-rw57h_7b9a072a-a040-414c-96c7-cf1454148745/horizon/0.log" Feb 20 09:09:33 crc kubenswrapper[4948]: I0220 09:09:33.443002 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw_c517042a-5354-4a8d-b7ff-c9aafa263b6c/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:09:33 crc kubenswrapper[4948]: I0220 09:09:33.554206 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-685c4db87b-rw57h_7b9a072a-a040-414c-96c7-cf1454148745/horizon-log/0.log" Feb 20 09:09:33 crc kubenswrapper[4948]: I0220 09:09:33.582627 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-4l55v_f42ddd0c-a2ec-414a-812d-f87213a3226d/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:09:33 crc kubenswrapper[4948]: I0220 09:09:33.745039 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29526301-gklw9_7990f62f-cd0f-4be2-b166-c00621d877e1/keystone-cron/0.log" Feb 20 09:09:33 crc kubenswrapper[4948]: I0220 09:09:33.859656 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6d4b8df464-wspg2_6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc/keystone-api/0.log" Feb 20 09:09:33 crc kubenswrapper[4948]: I0220 09:09:33.898533 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_24bf0fb4-81db-4955-8b5f-92c0f6b0856b/kube-state-metrics/0.log" Feb 20 09:09:34 crc kubenswrapper[4948]: I0220 09:09:34.062616 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs_6e619eac-dfdc-4c8c-aa51-fc08ba58b49a/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:09:34 crc kubenswrapper[4948]: I0220 09:09:34.381364 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-76b5684765-h5624_fb9d5a3e-b07f-4229-9b5b-d0db6750cf59/neutron-httpd/0.log" Feb 20 09:09:34 crc kubenswrapper[4948]: I0220 09:09:34.414673 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-76b5684765-h5624_fb9d5a3e-b07f-4229-9b5b-d0db6750cf59/neutron-api/0.log" Feb 20 09:09:34 crc kubenswrapper[4948]: I0220 09:09:34.472473 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr_fd280342-4276-4eee-a763-876e31be28c5/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:09:34 crc kubenswrapper[4948]: I0220 09:09:34.956432 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_354f9b2f-9eaa-4e47-b214-0fd895f90c4c/nova-api-log/0.log" Feb 20 09:09:35 crc kubenswrapper[4948]: I0220 09:09:35.175688 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_873a6553-9637-4f2a-a743-f33cde7fc883/nova-cell0-conductor-conductor/0.log" Feb 20 09:09:35 crc kubenswrapper[4948]: I0220 09:09:35.411454 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_e801ada9-0247-4b29-b262-04637e1f8452/nova-cell1-conductor-conductor/0.log" Feb 20 09:09:35 crc kubenswrapper[4948]: I0220 09:09:35.519498 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_9c00f24b-fedb-4747-91c4-3d0b551a7288/nova-cell1-novncproxy-novncproxy/0.log" Feb 20 09:09:35 crc kubenswrapper[4948]: I0220 09:09:35.531564 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_354f9b2f-9eaa-4e47-b214-0fd895f90c4c/nova-api-api/0.log" Feb 20 09:09:35 crc kubenswrapper[4948]: I0220 09:09:35.723683 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:09:35 crc kubenswrapper[4948]: E0220 09:09:35.723993 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:09:35 crc kubenswrapper[4948]: I0220 09:09:35.843604 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-p22xr_6005ded0-9e21-4908-bbf0-33c710ba4341/nova-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:09:36 crc kubenswrapper[4948]: I0220 09:09:36.083005 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_59965363-9971-4130-bb5a-5fcf5a44e2d5/nova-metadata-log/0.log" Feb 20 09:09:36 crc kubenswrapper[4948]: I0220 09:09:36.341176 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_d02114ae-21f2-4cc5-9bca-9505c25eaf5f/mysql-bootstrap/0.log" Feb 20 09:09:36 crc kubenswrapper[4948]: I0220 09:09:36.385566 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_f114bc41-01a5-4955-97eb-7fcf139cc5a9/nova-scheduler-scheduler/0.log" Feb 20 09:09:36 crc kubenswrapper[4948]: I0220 09:09:36.509923 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_d02114ae-21f2-4cc5-9bca-9505c25eaf5f/mysql-bootstrap/0.log" Feb 20 09:09:36 crc kubenswrapper[4948]: I0220 09:09:36.539391 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_d02114ae-21f2-4cc5-9bca-9505c25eaf5f/galera/0.log" Feb 20 09:09:36 crc kubenswrapper[4948]: I0220 09:09:36.757428 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_6e59b63d-0c56-488c-87cb-348af87058c6/mysql-bootstrap/0.log" Feb 20 09:09:36 crc kubenswrapper[4948]: I0220 09:09:36.957406 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_6e59b63d-0c56-488c-87cb-348af87058c6/mysql-bootstrap/0.log" Feb 20 09:09:36 crc kubenswrapper[4948]: I0220 09:09:36.975160 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_6e59b63d-0c56-488c-87cb-348af87058c6/galera/0.log" Feb 20 09:09:37 crc kubenswrapper[4948]: I0220 09:09:37.128538 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_b2a76f68-3b4b-4993-aa4f-3b31b4ee1008/openstackclient/0.log" Feb 20 09:09:37 crc kubenswrapper[4948]: I0220 09:09:37.259184 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-gcfn8_1510e8b2-af6f-4641-a2d2-361830382250/openstack-network-exporter/0.log" Feb 20 09:09:37 crc kubenswrapper[4948]: I0220 09:09:37.301333 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_59965363-9971-4130-bb5a-5fcf5a44e2d5/nova-metadata-metadata/0.log" Feb 20 09:09:37 crc kubenswrapper[4948]: I0220 09:09:37.477497 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-mkpng_b80cb988-de2a-4e65-9161-a0af0561c754/ovn-controller/0.log" Feb 20 09:09:37 crc kubenswrapper[4948]: I0220 09:09:37.501379 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-q4bv4_c82772b6-9737-45ee-998a-e897086c03b5/ovsdb-server-init/0.log" Feb 20 09:09:37 crc kubenswrapper[4948]: I0220 09:09:37.691377 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-q4bv4_c82772b6-9737-45ee-998a-e897086c03b5/ovsdb-server-init/0.log" Feb 20 09:09:37 crc kubenswrapper[4948]: I0220 09:09:37.694023 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-q4bv4_c82772b6-9737-45ee-998a-e897086c03b5/ovs-vswitchd/0.log" Feb 20 09:09:37 crc kubenswrapper[4948]: I0220 09:09:37.710111 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-q4bv4_c82772b6-9737-45ee-998a-e897086c03b5/ovsdb-server/0.log" Feb 20 09:09:37 crc kubenswrapper[4948]: I0220 09:09:37.950432 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-4b2vp_b19ca7cb-01d2-4965-90f6-5d50211870cc/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:09:37 crc kubenswrapper[4948]: I0220 09:09:37.988789 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_eeb3782c-0a88-4d17-b481-e2aef7ea386a/openstack-network-exporter/0.log" Feb 20 09:09:38 crc kubenswrapper[4948]: I0220 09:09:38.127300 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_eeb3782c-0a88-4d17-b481-e2aef7ea386a/ovn-northd/0.log" Feb 20 09:09:38 crc kubenswrapper[4948]: I0220 09:09:38.153530 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_d681442e-6bdb-46e4-9b49-6466e5e036a0/openstack-network-exporter/0.log" Feb 20 09:09:38 crc kubenswrapper[4948]: I0220 09:09:38.193936 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_d681442e-6bdb-46e4-9b49-6466e5e036a0/ovsdbserver-nb/0.log" Feb 20 09:09:38 crc kubenswrapper[4948]: I0220 09:09:38.344135 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_4b04e19d-cd3d-474d-884d-df12d0e1fffd/openstack-network-exporter/0.log" Feb 20 09:09:38 crc kubenswrapper[4948]: I0220 09:09:38.450488 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_4b04e19d-cd3d-474d-884d-df12d0e1fffd/ovsdbserver-sb/0.log" Feb 20 09:09:38 crc kubenswrapper[4948]: I0220 09:09:38.577387 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-86887846d-pqjcb_5047ff00-845b-459f-ac89-6cbf997f93fb/placement-api/0.log" Feb 20 09:09:38 crc kubenswrapper[4948]: I0220 09:09:38.662231 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-86887846d-pqjcb_5047ff00-845b-459f-ac89-6cbf997f93fb/placement-log/0.log" Feb 20 09:09:38 crc kubenswrapper[4948]: I0220 09:09:38.666536 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_037b329f-712f-4a67-984a-75affd2a57b7/setup-container/0.log" Feb 20 09:09:38 crc kubenswrapper[4948]: I0220 09:09:38.954844 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_037b329f-712f-4a67-984a-75affd2a57b7/setup-container/0.log" Feb 20 09:09:38 crc kubenswrapper[4948]: I0220 09:09:38.986918 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_037b329f-712f-4a67-984a-75affd2a57b7/rabbitmq/0.log" Feb 20 09:09:39 crc kubenswrapper[4948]: I0220 09:09:39.173303 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f063bb62-10c3-46d9-a41d-56d2826059e0/setup-container/0.log" Feb 20 09:09:39 crc kubenswrapper[4948]: I0220 09:09:39.300011 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f063bb62-10c3-46d9-a41d-56d2826059e0/setup-container/0.log" Feb 20 09:09:39 crc kubenswrapper[4948]: I0220 09:09:39.391738 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f063bb62-10c3-46d9-a41d-56d2826059e0/rabbitmq/0.log" Feb 20 09:09:39 crc kubenswrapper[4948]: I0220 09:09:39.485199 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-ds522_e3b05342-f970-4fdb-9b1c-ff521e118f96/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:09:39 crc kubenswrapper[4948]: I0220 09:09:39.568728 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-f7ww9_c3e7cce4-5144-4190-8013-83428c66cde9/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:09:39 crc kubenswrapper[4948]: I0220 09:09:39.700632 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch_90b6e7df-b3e1-4d05-a6b3-7fe47d70804f/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:09:39 crc kubenswrapper[4948]: I0220 09:09:39.800522 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-lkbkf_79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:09:40 crc kubenswrapper[4948]: I0220 09:09:40.009594 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-cqchc_41856163-b621-439a-95c1-ca8d13f0a08c/ssh-known-hosts-edpm-deployment/0.log" Feb 20 09:09:40 crc kubenswrapper[4948]: I0220 09:09:40.172553 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7c7cd9876c-262qf_30f48966-e44d-44bc-a868-7e8119ac186c/proxy-httpd/0.log" Feb 20 09:09:40 crc kubenswrapper[4948]: I0220 09:09:40.225120 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-dllls_642adfd1-e253-4e0a-80e6-860a9f5c1ae1/swift-ring-rebalance/0.log" Feb 20 09:09:40 crc kubenswrapper[4948]: I0220 09:09:40.228202 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7c7cd9876c-262qf_30f48966-e44d-44bc-a868-7e8119ac186c/proxy-server/0.log" Feb 20 09:09:40 crc kubenswrapper[4948]: I0220 09:09:40.410742 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/account-reaper/0.log" Feb 20 09:09:40 crc kubenswrapper[4948]: I0220 09:09:40.420955 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/account-auditor/0.log" Feb 20 09:09:40 crc kubenswrapper[4948]: I0220 09:09:40.467643 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/account-replicator/0.log" Feb 20 09:09:40 crc kubenswrapper[4948]: I0220 09:09:40.608333 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/container-auditor/0.log" Feb 20 09:09:40 crc kubenswrapper[4948]: I0220 09:09:40.612558 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/account-server/0.log" Feb 20 09:09:40 crc kubenswrapper[4948]: I0220 09:09:40.672175 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/container-server/0.log" Feb 20 09:09:40 crc kubenswrapper[4948]: I0220 09:09:40.674237 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/container-replicator/0.log" Feb 20 09:09:40 crc kubenswrapper[4948]: I0220 09:09:40.783919 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/container-updater/0.log" Feb 20 09:09:40 crc kubenswrapper[4948]: I0220 09:09:40.816808 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/object-auditor/0.log" Feb 20 09:09:40 crc kubenswrapper[4948]: I0220 09:09:40.898573 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/object-expirer/0.log" Feb 20 09:09:40 crc kubenswrapper[4948]: I0220 09:09:40.908554 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/object-replicator/0.log" Feb 20 09:09:41 crc kubenswrapper[4948]: I0220 09:09:41.021593 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/object-updater/0.log" Feb 20 09:09:41 crc kubenswrapper[4948]: I0220 09:09:41.026903 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/object-server/0.log" Feb 20 09:09:41 crc kubenswrapper[4948]: I0220 09:09:41.101100 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/rsync/0.log" Feb 20 09:09:41 crc kubenswrapper[4948]: I0220 09:09:41.106834 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/swift-recon-cron/0.log" Feb 20 09:09:41 crc kubenswrapper[4948]: I0220 09:09:41.339892 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd_06ed5839-bc31-4691-8586-cd42c0413006/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:09:41 crc kubenswrapper[4948]: I0220 09:09:41.358128 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_0fefa0c4-ece9-4caf-8b41-8a64472c53e6/tempest-tests-tempest-tests-runner/0.log" Feb 20 09:09:41 crc kubenswrapper[4948]: I0220 09:09:41.507949 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_de083766-efe0-42cd-95c5-b87bbeb33b0a/test-operator-logs-container/0.log" Feb 20 09:09:41 crc kubenswrapper[4948]: I0220 09:09:41.611710 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-lf98d_2e05be38-f4ab-415f-a71d-9e5233cf0ea7/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:09:50 crc kubenswrapper[4948]: I0220 09:09:50.722192 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:09:50 crc kubenswrapper[4948]: E0220 09:09:50.723908 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:09:52 crc kubenswrapper[4948]: I0220 09:09:52.316865 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_8e994489-a2cf-4f6d-a00c-98f627ba0e5f/memcached/0.log" Feb 20 09:10:04 crc kubenswrapper[4948]: I0220 09:10:04.722494 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:10:04 crc kubenswrapper[4948]: E0220 09:10:04.723373 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:10:07 crc kubenswrapper[4948]: I0220 09:10:07.307269 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6_a1b99eb4-2ab2-4755-92ed-a5226d91bfaf/util/0.log" Feb 20 09:10:07 crc kubenswrapper[4948]: I0220 09:10:07.510607 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6_a1b99eb4-2ab2-4755-92ed-a5226d91bfaf/pull/0.log" Feb 20 09:10:07 crc kubenswrapper[4948]: I0220 09:10:07.514336 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6_a1b99eb4-2ab2-4755-92ed-a5226d91bfaf/util/0.log" Feb 20 09:10:07 crc kubenswrapper[4948]: I0220 09:10:07.529193 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6_a1b99eb4-2ab2-4755-92ed-a5226d91bfaf/pull/0.log" Feb 20 09:10:07 crc kubenswrapper[4948]: I0220 09:10:07.694107 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6_a1b99eb4-2ab2-4755-92ed-a5226d91bfaf/extract/0.log" Feb 20 09:10:07 crc kubenswrapper[4948]: I0220 09:10:07.714485 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6_a1b99eb4-2ab2-4755-92ed-a5226d91bfaf/pull/0.log" Feb 20 09:10:07 crc kubenswrapper[4948]: I0220 09:10:07.719090 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6_a1b99eb4-2ab2-4755-92ed-a5226d91bfaf/util/0.log" Feb 20 09:10:08 crc kubenswrapper[4948]: I0220 09:10:08.232588 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-6d8bf5c495-kjpkr_98a1aa06-948b-4034-bc07-7e546e341a8f/manager/0.log" Feb 20 09:10:08 crc kubenswrapper[4948]: I0220 09:10:08.492153 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987464f4-xc74s_347cdd7c-e5e8-49b1-a4f1-687b8a06b250/manager/0.log" Feb 20 09:10:08 crc kubenswrapper[4948]: I0220 09:10:08.659516 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-69f49c598c-6cr5p_2d303d1a-1062-4bf6-be68-1bd6d1a3228f/manager/0.log" Feb 20 09:10:08 crc kubenswrapper[4948]: I0220 09:10:08.901883 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5b9b8895d5-bkl49_4a663826-d4b5-4ed3-8270-099b003390b4/manager/0.log" Feb 20 09:10:09 crc kubenswrapper[4948]: I0220 09:10:09.192436 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-5d946d989d-c6xqd_586113f2-38b6-4bd1-8adb-3c155bb35ba9/manager/0.log" Feb 20 09:10:09 crc kubenswrapper[4948]: I0220 09:10:09.366201 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-554564d7fc-5fqzw_105a118c-121d-4582-960f-1da9957980cb/manager/0.log" Feb 20 09:10:09 crc kubenswrapper[4948]: I0220 09:10:09.414728 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-79d975b745-jrlw5_93b93501-acfe-4274-9a22-ca644b1d11d3/manager/0.log" Feb 20 09:10:09 crc kubenswrapper[4948]: I0220 09:10:09.664120 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-b4d948c87-ttkfm_3fbbd11d-612d-479a-a34f-505d995a4871/manager/0.log" Feb 20 09:10:09 crc kubenswrapper[4948]: I0220 09:10:09.667161 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-54f6768c69-lj2sz_f762ec0d-e09c-4baf-9540-f4ac61ce7234/manager/0.log" Feb 20 09:10:09 crc kubenswrapper[4948]: I0220 09:10:09.851923 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6994f66f48-zcccd_6e8e8b87-5a00-486f-a00c-1450c109f3b2/manager/0.log" Feb 20 09:10:10 crc kubenswrapper[4948]: I0220 09:10:10.388929 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-64ddbf8bb-jcnxd_88142137-864d-4660-a688-a7dcc503851b/manager/0.log" Feb 20 09:10:10 crc kubenswrapper[4948]: I0220 09:10:10.456535 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-567668f5cf-jxwsx_7c7bb531-7900-4cc6-9d9b-bae52dabc59d/manager/0.log" Feb 20 09:10:10 crc kubenswrapper[4948]: I0220 09:10:10.699524 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m_b8b07685-095c-4bbd-a30d-57ca59d7cbdc/manager/0.log" Feb 20 09:10:11 crc kubenswrapper[4948]: I0220 09:10:11.021034 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-59c446d8d6-4cs2n_9465a7a8-c08e-44c9-a76d-3878b0bff3af/operator/0.log" Feb 20 09:10:11 crc kubenswrapper[4948]: I0220 09:10:11.306257 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-9z4jj_64872326-fb95-4bc1-a6c5-0b34242883f9/registry-server/0.log" Feb 20 09:10:11 crc kubenswrapper[4948]: I0220 09:10:11.560285 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-d44cf6b75-dkkq9_3ea675de-d1b0-4880-9652-eb066f6b0fb7/manager/0.log" Feb 20 09:10:11 crc kubenswrapper[4948]: I0220 09:10:11.799746 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-8497b45c89-znrkz_16933747-642c-45ff-9f98-9321c633826a/manager/0.log" Feb 20 09:10:11 crc kubenswrapper[4948]: I0220 09:10:11.946899 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69f8888797-hl5w7_ee33bd02-a3ce-415d-9d2d-fefd383e9810/manager/0.log" Feb 20 09:10:12 crc kubenswrapper[4948]: I0220 09:10:12.004503 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-59pcs_fe0aca33-1acf-463a-91a6-bbf35a38fd7c/operator/0.log" Feb 20 09:10:12 crc kubenswrapper[4948]: I0220 09:10:12.183362 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-68f46476f-5lfts_d65e0993-11a0-4e81-963f-eeb9dcb92536/manager/0.log" Feb 20 09:10:12 crc kubenswrapper[4948]: I0220 09:10:12.297104 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7f45b4ff68-5rrd9_0f0cbb65-3c8e-41e6-8059-34e121de0821/manager/0.log" Feb 20 09:10:12 crc kubenswrapper[4948]: I0220 09:10:12.363796 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-7866795846-f7cxn_0f98bc58-d0e3-405b-88fd-d8bd65f415a4/manager/0.log" Feb 20 09:10:12 crc kubenswrapper[4948]: I0220 09:10:12.577549 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-5db88f68c-pndq7_f2400173-aa87-476c-8216-4f8c9cf9d474/manager/0.log" Feb 20 09:10:12 crc kubenswrapper[4948]: I0220 09:10:12.808701 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-8569bc6fdb-lt4jk_e9f13bcd-ac5a-4cbd-952a-a5b9dceee562/manager/0.log" Feb 20 09:10:14 crc kubenswrapper[4948]: I0220 09:10:14.443678 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-868647ff47-4h9v5_b8e12d0c-9564-4bf2-ac61-b22d2fbdf855/manager/0.log" Feb 20 09:10:19 crc kubenswrapper[4948]: I0220 09:10:19.723158 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:10:20 crc kubenswrapper[4948]: I0220 09:10:20.814970 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"5bbfc48505968074b16a56aea08c035f5c0687419ea24b66d69a44ebd2ab9d7d"} Feb 20 09:10:31 crc kubenswrapper[4948]: I0220 09:10:31.561016 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-bgm85_a149fe37-c748-4120-9116-1da4b680d880/control-plane-machine-set-operator/0.log" Feb 20 09:10:31 crc kubenswrapper[4948]: I0220 09:10:31.717782 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-fthdq_5bc3d0cd-eacd-4b25-9acd-853e49db7b47/kube-rbac-proxy/0.log" Feb 20 09:10:31 crc kubenswrapper[4948]: I0220 09:10:31.718723 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-fthdq_5bc3d0cd-eacd-4b25-9acd-853e49db7b47/machine-api-operator/0.log" Feb 20 09:10:45 crc kubenswrapper[4948]: I0220 09:10:45.481761 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-smtqw_1437d818-abf4-4602-8a28-e88a76e482a9/cert-manager-controller/0.log" Feb 20 09:10:45 crc kubenswrapper[4948]: I0220 09:10:45.654120 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-9llfv_bb56399a-a840-49af-972f-4f32c91efed7/cert-manager-cainjector/0.log" Feb 20 09:10:45 crc kubenswrapper[4948]: I0220 09:10:45.711084 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-f4bjc_305a17f4-aef3-4036-8fce-3756ff5bbd2f/cert-manager-webhook/0.log" Feb 20 09:10:58 crc kubenswrapper[4948]: I0220 09:10:58.488917 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5c78fc5d65-9mv9z_47461ed8-c51b-4014-bcf0-4d95a0278f85/nmstate-console-plugin/0.log" Feb 20 09:10:58 crc kubenswrapper[4948]: I0220 09:10:58.608757 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-dxdh6_3a869b3a-b757-49ba-9096-2e562d980aae/nmstate-handler/0.log" Feb 20 09:10:58 crc kubenswrapper[4948]: I0220 09:10:58.666656 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58c85c668d-tdst7_a7b34c06-d0c5-4f20-9521-eec500ea2510/nmstate-metrics/0.log" Feb 20 09:10:58 crc kubenswrapper[4948]: I0220 09:10:58.692412 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58c85c668d-tdst7_a7b34c06-d0c5-4f20-9521-eec500ea2510/kube-rbac-proxy/0.log" Feb 20 09:10:58 crc kubenswrapper[4948]: I0220 09:10:58.816068 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-694c9596b7-xqtvg_59277281-0112-45da-a64a-1b1d02b90473/nmstate-operator/0.log" Feb 20 09:10:58 crc kubenswrapper[4948]: I0220 09:10:58.861625 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-866bcb46dc-77sdv_da0e4284-f20d-44a8-b529-e388f15822d3/nmstate-webhook/0.log" Feb 20 09:11:27 crc kubenswrapper[4948]: I0220 09:11:27.705865 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-c6sqk_68ee0185-13d4-45c7-9a39-bb474e9314d0/kube-rbac-proxy/0.log" Feb 20 09:11:27 crc kubenswrapper[4948]: I0220 09:11:27.783424 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-c6sqk_68ee0185-13d4-45c7-9a39-bb474e9314d0/controller/0.log" Feb 20 09:11:27 crc kubenswrapper[4948]: I0220 09:11:27.959565 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-frr-files/0.log" Feb 20 09:11:28 crc kubenswrapper[4948]: I0220 09:11:28.134463 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-frr-files/0.log" Feb 20 09:11:28 crc kubenswrapper[4948]: I0220 09:11:28.164319 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-metrics/0.log" Feb 20 09:11:28 crc kubenswrapper[4948]: I0220 09:11:28.169796 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-reloader/0.log" Feb 20 09:11:28 crc kubenswrapper[4948]: I0220 09:11:28.215301 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-reloader/0.log" Feb 20 09:11:28 crc kubenswrapper[4948]: I0220 09:11:28.399921 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-frr-files/0.log" Feb 20 09:11:28 crc kubenswrapper[4948]: I0220 09:11:28.408036 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-metrics/0.log" Feb 20 09:11:28 crc kubenswrapper[4948]: I0220 09:11:28.422811 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-reloader/0.log" Feb 20 09:11:28 crc kubenswrapper[4948]: I0220 09:11:28.440364 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-metrics/0.log" Feb 20 09:11:28 crc kubenswrapper[4948]: I0220 09:11:28.605323 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-frr-files/0.log" Feb 20 09:11:28 crc kubenswrapper[4948]: I0220 09:11:28.633253 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-reloader/0.log" Feb 20 09:11:28 crc kubenswrapper[4948]: I0220 09:11:28.658694 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/controller/0.log" Feb 20 09:11:28 crc kubenswrapper[4948]: I0220 09:11:28.660088 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-metrics/0.log" Feb 20 09:11:28 crc kubenswrapper[4948]: I0220 09:11:28.837670 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/kube-rbac-proxy-frr/0.log" Feb 20 09:11:28 crc kubenswrapper[4948]: I0220 09:11:28.857443 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/frr-metrics/0.log" Feb 20 09:11:28 crc kubenswrapper[4948]: I0220 09:11:28.887725 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/kube-rbac-proxy/0.log" Feb 20 09:11:29 crc kubenswrapper[4948]: I0220 09:11:29.016472 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/reloader/0.log" Feb 20 09:11:29 crc kubenswrapper[4948]: I0220 09:11:29.132898 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-78b44bf5bb-7zsxs_2506d4e2-b434-4fe4-970e-7cd14601677d/frr-k8s-webhook-server/0.log" Feb 20 09:11:29 crc kubenswrapper[4948]: I0220 09:11:29.371889 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-8478c6b8cc-v98zs_b4f86a2e-15f9-441c-953b-49f331d4122e/manager/0.log" Feb 20 09:11:29 crc kubenswrapper[4948]: I0220 09:11:29.440844 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5fcd7d546c-kgpxs_07022b13-7dc1-45d1-a626-91792bd3aa90/webhook-server/0.log" Feb 20 09:11:29 crc kubenswrapper[4948]: I0220 09:11:29.639730 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-h7xhj_2c199acd-2fd3-4b9f-b50c-a266191c4777/kube-rbac-proxy/0.log" Feb 20 09:11:30 crc kubenswrapper[4948]: I0220 09:11:30.139988 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-h7xhj_2c199acd-2fd3-4b9f-b50c-a266191c4777/speaker/0.log" Feb 20 09:11:30 crc kubenswrapper[4948]: I0220 09:11:30.246447 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/frr/0.log" Feb 20 09:11:43 crc kubenswrapper[4948]: I0220 09:11:43.207899 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct_9e5e2f81-8ea9-4ddb-b2ab-584c40b15916/util/0.log" Feb 20 09:11:43 crc kubenswrapper[4948]: I0220 09:11:43.382877 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct_9e5e2f81-8ea9-4ddb-b2ab-584c40b15916/util/0.log" Feb 20 09:11:43 crc kubenswrapper[4948]: I0220 09:11:43.435636 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct_9e5e2f81-8ea9-4ddb-b2ab-584c40b15916/pull/0.log" Feb 20 09:11:43 crc kubenswrapper[4948]: I0220 09:11:43.474397 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct_9e5e2f81-8ea9-4ddb-b2ab-584c40b15916/pull/0.log" Feb 20 09:11:43 crc kubenswrapper[4948]: I0220 09:11:43.570367 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct_9e5e2f81-8ea9-4ddb-b2ab-584c40b15916/util/0.log" Feb 20 09:11:43 crc kubenswrapper[4948]: I0220 09:11:43.598168 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct_9e5e2f81-8ea9-4ddb-b2ab-584c40b15916/pull/0.log" Feb 20 09:11:43 crc kubenswrapper[4948]: I0220 09:11:43.635788 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct_9e5e2f81-8ea9-4ddb-b2ab-584c40b15916/extract/0.log" Feb 20 09:11:43 crc kubenswrapper[4948]: I0220 09:11:43.778226 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tt6t6_94e3bb0b-818a-491e-8e1e-a1664f3a9f12/extract-utilities/0.log" Feb 20 09:11:43 crc kubenswrapper[4948]: I0220 09:11:43.971657 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tt6t6_94e3bb0b-818a-491e-8e1e-a1664f3a9f12/extract-utilities/0.log" Feb 20 09:11:43 crc kubenswrapper[4948]: I0220 09:11:43.971682 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tt6t6_94e3bb0b-818a-491e-8e1e-a1664f3a9f12/extract-content/0.log" Feb 20 09:11:43 crc kubenswrapper[4948]: I0220 09:11:43.999182 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tt6t6_94e3bb0b-818a-491e-8e1e-a1664f3a9f12/extract-content/0.log" Feb 20 09:11:44 crc kubenswrapper[4948]: I0220 09:11:44.158733 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tt6t6_94e3bb0b-818a-491e-8e1e-a1664f3a9f12/extract-content/0.log" Feb 20 09:11:44 crc kubenswrapper[4948]: I0220 09:11:44.174505 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tt6t6_94e3bb0b-818a-491e-8e1e-a1664f3a9f12/extract-utilities/0.log" Feb 20 09:11:44 crc kubenswrapper[4948]: I0220 09:11:44.351847 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dl7nj_0e759dc6-36e8-4a97-b994-e657f1b5d540/extract-utilities/0.log" Feb 20 09:11:44 crc kubenswrapper[4948]: I0220 09:11:44.635595 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dl7nj_0e759dc6-36e8-4a97-b994-e657f1b5d540/extract-utilities/0.log" Feb 20 09:11:44 crc kubenswrapper[4948]: I0220 09:11:44.638998 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dl7nj_0e759dc6-36e8-4a97-b994-e657f1b5d540/extract-content/0.log" Feb 20 09:11:44 crc kubenswrapper[4948]: I0220 09:11:44.729730 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dl7nj_0e759dc6-36e8-4a97-b994-e657f1b5d540/extract-content/0.log" Feb 20 09:11:44 crc kubenswrapper[4948]: I0220 09:11:44.756828 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tt6t6_94e3bb0b-818a-491e-8e1e-a1664f3a9f12/registry-server/0.log" Feb 20 09:11:44 crc kubenswrapper[4948]: I0220 09:11:44.869208 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dl7nj_0e759dc6-36e8-4a97-b994-e657f1b5d540/extract-utilities/0.log" Feb 20 09:11:44 crc kubenswrapper[4948]: I0220 09:11:44.911955 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dl7nj_0e759dc6-36e8-4a97-b994-e657f1b5d540/extract-content/0.log" Feb 20 09:11:45 crc kubenswrapper[4948]: I0220 09:11:45.074101 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w_bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a/util/0.log" Feb 20 09:11:45 crc kubenswrapper[4948]: I0220 09:11:45.334627 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w_bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a/pull/0.log" Feb 20 09:11:45 crc kubenswrapper[4948]: I0220 09:11:45.344033 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w_bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a/util/0.log" Feb 20 09:11:45 crc kubenswrapper[4948]: I0220 09:11:45.402411 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dl7nj_0e759dc6-36e8-4a97-b994-e657f1b5d540/registry-server/0.log" Feb 20 09:11:45 crc kubenswrapper[4948]: I0220 09:11:45.408240 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w_bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a/pull/0.log" Feb 20 09:11:45 crc kubenswrapper[4948]: I0220 09:11:45.507856 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w_bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a/util/0.log" Feb 20 09:11:45 crc kubenswrapper[4948]: I0220 09:11:45.551238 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w_bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a/pull/0.log" Feb 20 09:11:45 crc kubenswrapper[4948]: I0220 09:11:45.584479 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w_bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a/extract/0.log" Feb 20 09:11:45 crc kubenswrapper[4948]: I0220 09:11:45.744738 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x7bcm_6ec98112-e189-4a27-a6dd-8cb035596810/extract-utilities/0.log" Feb 20 09:11:45 crc kubenswrapper[4948]: I0220 09:11:45.777277 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-tbqzs_1e3d2cd4-4311-4454-b257-20a3caf243b3/marketplace-operator/0.log" Feb 20 09:11:45 crc kubenswrapper[4948]: I0220 09:11:45.944292 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x7bcm_6ec98112-e189-4a27-a6dd-8cb035596810/extract-content/0.log" Feb 20 09:11:45 crc kubenswrapper[4948]: I0220 09:11:45.974712 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x7bcm_6ec98112-e189-4a27-a6dd-8cb035596810/extract-content/0.log" Feb 20 09:11:45 crc kubenswrapper[4948]: I0220 09:11:45.981486 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x7bcm_6ec98112-e189-4a27-a6dd-8cb035596810/extract-utilities/0.log" Feb 20 09:11:46 crc kubenswrapper[4948]: I0220 09:11:46.123896 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x7bcm_6ec98112-e189-4a27-a6dd-8cb035596810/extract-utilities/0.log" Feb 20 09:11:46 crc kubenswrapper[4948]: I0220 09:11:46.215445 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x7bcm_6ec98112-e189-4a27-a6dd-8cb035596810/extract-content/0.log" Feb 20 09:11:46 crc kubenswrapper[4948]: I0220 09:11:46.283814 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x7bcm_6ec98112-e189-4a27-a6dd-8cb035596810/registry-server/0.log" Feb 20 09:11:46 crc kubenswrapper[4948]: I0220 09:11:46.332439 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jbtm5_5ee242c9-9783-44cb-a4dc-1ea50a7ae185/extract-utilities/0.log" Feb 20 09:11:46 crc kubenswrapper[4948]: I0220 09:11:46.447260 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jbtm5_5ee242c9-9783-44cb-a4dc-1ea50a7ae185/extract-content/0.log" Feb 20 09:11:46 crc kubenswrapper[4948]: I0220 09:11:46.484618 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jbtm5_5ee242c9-9783-44cb-a4dc-1ea50a7ae185/extract-content/0.log" Feb 20 09:11:46 crc kubenswrapper[4948]: I0220 09:11:46.504658 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jbtm5_5ee242c9-9783-44cb-a4dc-1ea50a7ae185/extract-utilities/0.log" Feb 20 09:11:46 crc kubenswrapper[4948]: I0220 09:11:46.670905 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jbtm5_5ee242c9-9783-44cb-a4dc-1ea50a7ae185/extract-utilities/0.log" Feb 20 09:11:46 crc kubenswrapper[4948]: I0220 09:11:46.681053 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jbtm5_5ee242c9-9783-44cb-a4dc-1ea50a7ae185/extract-content/0.log" Feb 20 09:11:47 crc kubenswrapper[4948]: I0220 09:11:47.130722 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jbtm5_5ee242c9-9783-44cb-a4dc-1ea50a7ae185/registry-server/0.log" Feb 20 09:11:48 crc kubenswrapper[4948]: I0220 09:11:48.034608 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7sjmh"] Feb 20 09:11:48 crc kubenswrapper[4948]: E0220 09:11:48.035155 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cfde212-7150-4b57-b3dd-41715c736426" containerName="container-00" Feb 20 09:11:48 crc kubenswrapper[4948]: I0220 09:11:48.035176 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cfde212-7150-4b57-b3dd-41715c736426" containerName="container-00" Feb 20 09:11:48 crc kubenswrapper[4948]: I0220 09:11:48.035502 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cfde212-7150-4b57-b3dd-41715c736426" containerName="container-00" Feb 20 09:11:48 crc kubenswrapper[4948]: I0220 09:11:48.037238 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7sjmh" Feb 20 09:11:48 crc kubenswrapper[4948]: I0220 09:11:48.063773 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7sjmh"] Feb 20 09:11:48 crc kubenswrapper[4948]: I0220 09:11:48.113642 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4v6l\" (UniqueName: \"kubernetes.io/projected/654b30bf-3b65-4501-9a0e-9ebc33812c73-kube-api-access-g4v6l\") pod \"redhat-marketplace-7sjmh\" (UID: \"654b30bf-3b65-4501-9a0e-9ebc33812c73\") " pod="openshift-marketplace/redhat-marketplace-7sjmh" Feb 20 09:11:48 crc kubenswrapper[4948]: I0220 09:11:48.113734 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/654b30bf-3b65-4501-9a0e-9ebc33812c73-utilities\") pod \"redhat-marketplace-7sjmh\" (UID: \"654b30bf-3b65-4501-9a0e-9ebc33812c73\") " pod="openshift-marketplace/redhat-marketplace-7sjmh" Feb 20 09:11:48 crc kubenswrapper[4948]: I0220 09:11:48.113900 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/654b30bf-3b65-4501-9a0e-9ebc33812c73-catalog-content\") pod \"redhat-marketplace-7sjmh\" (UID: \"654b30bf-3b65-4501-9a0e-9ebc33812c73\") " pod="openshift-marketplace/redhat-marketplace-7sjmh" Feb 20 09:11:48 crc kubenswrapper[4948]: I0220 09:11:48.215375 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/654b30bf-3b65-4501-9a0e-9ebc33812c73-catalog-content\") pod \"redhat-marketplace-7sjmh\" (UID: \"654b30bf-3b65-4501-9a0e-9ebc33812c73\") " pod="openshift-marketplace/redhat-marketplace-7sjmh" Feb 20 09:11:48 crc kubenswrapper[4948]: I0220 09:11:48.215621 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4v6l\" (UniqueName: \"kubernetes.io/projected/654b30bf-3b65-4501-9a0e-9ebc33812c73-kube-api-access-g4v6l\") pod \"redhat-marketplace-7sjmh\" (UID: \"654b30bf-3b65-4501-9a0e-9ebc33812c73\") " pod="openshift-marketplace/redhat-marketplace-7sjmh" Feb 20 09:11:48 crc kubenswrapper[4948]: I0220 09:11:48.215655 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/654b30bf-3b65-4501-9a0e-9ebc33812c73-utilities\") pod \"redhat-marketplace-7sjmh\" (UID: \"654b30bf-3b65-4501-9a0e-9ebc33812c73\") " pod="openshift-marketplace/redhat-marketplace-7sjmh" Feb 20 09:11:48 crc kubenswrapper[4948]: I0220 09:11:48.216077 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/654b30bf-3b65-4501-9a0e-9ebc33812c73-catalog-content\") pod \"redhat-marketplace-7sjmh\" (UID: \"654b30bf-3b65-4501-9a0e-9ebc33812c73\") " pod="openshift-marketplace/redhat-marketplace-7sjmh" Feb 20 09:11:48 crc kubenswrapper[4948]: I0220 09:11:48.216153 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/654b30bf-3b65-4501-9a0e-9ebc33812c73-utilities\") pod \"redhat-marketplace-7sjmh\" (UID: \"654b30bf-3b65-4501-9a0e-9ebc33812c73\") " pod="openshift-marketplace/redhat-marketplace-7sjmh" Feb 20 09:11:48 crc kubenswrapper[4948]: I0220 09:11:48.234935 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4v6l\" (UniqueName: \"kubernetes.io/projected/654b30bf-3b65-4501-9a0e-9ebc33812c73-kube-api-access-g4v6l\") pod \"redhat-marketplace-7sjmh\" (UID: \"654b30bf-3b65-4501-9a0e-9ebc33812c73\") " pod="openshift-marketplace/redhat-marketplace-7sjmh" Feb 20 09:11:48 crc kubenswrapper[4948]: I0220 09:11:48.365315 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7sjmh" Feb 20 09:11:48 crc kubenswrapper[4948]: I0220 09:11:48.833940 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7sjmh"] Feb 20 09:11:49 crc kubenswrapper[4948]: I0220 09:11:49.618245 4948 generic.go:334] "Generic (PLEG): container finished" podID="654b30bf-3b65-4501-9a0e-9ebc33812c73" containerID="0498136ca7009d716fd78a1af98d3ae2091af6957f0aa8b1253de62c66c87915" exitCode=0 Feb 20 09:11:49 crc kubenswrapper[4948]: I0220 09:11:49.618304 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sjmh" event={"ID":"654b30bf-3b65-4501-9a0e-9ebc33812c73","Type":"ContainerDied","Data":"0498136ca7009d716fd78a1af98d3ae2091af6957f0aa8b1253de62c66c87915"} Feb 20 09:11:49 crc kubenswrapper[4948]: I0220 09:11:49.618335 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sjmh" event={"ID":"654b30bf-3b65-4501-9a0e-9ebc33812c73","Type":"ContainerStarted","Data":"935397e9a001171af02db7a3be1a28e8bc26a789185e6481f73a103b9280bae0"} Feb 20 09:11:49 crc kubenswrapper[4948]: I0220 09:11:49.621397 4948 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 09:11:50 crc kubenswrapper[4948]: I0220 09:11:50.632684 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sjmh" event={"ID":"654b30bf-3b65-4501-9a0e-9ebc33812c73","Type":"ContainerStarted","Data":"f78e60617dfa8cf4a04999e07015aa7e834279c7e0b99a1ff6c2ec26cb91d973"} Feb 20 09:11:51 crc kubenswrapper[4948]: I0220 09:11:51.644766 4948 generic.go:334] "Generic (PLEG): container finished" podID="654b30bf-3b65-4501-9a0e-9ebc33812c73" containerID="f78e60617dfa8cf4a04999e07015aa7e834279c7e0b99a1ff6c2ec26cb91d973" exitCode=0 Feb 20 09:11:51 crc kubenswrapper[4948]: I0220 09:11:51.644830 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sjmh" event={"ID":"654b30bf-3b65-4501-9a0e-9ebc33812c73","Type":"ContainerDied","Data":"f78e60617dfa8cf4a04999e07015aa7e834279c7e0b99a1ff6c2ec26cb91d973"} Feb 20 09:11:52 crc kubenswrapper[4948]: I0220 09:11:52.655645 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sjmh" event={"ID":"654b30bf-3b65-4501-9a0e-9ebc33812c73","Type":"ContainerStarted","Data":"e00180d9024f5ed42ba57eca358c05e3488371a087ae601107ee28e8c4a0c9d7"} Feb 20 09:11:52 crc kubenswrapper[4948]: I0220 09:11:52.683021 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7sjmh" podStartSLOduration=2.273941058 podStartE2EDuration="4.682959414s" podCreationTimestamp="2026-02-20 09:11:48 +0000 UTC" firstStartedPulling="2026-02-20 09:11:49.621121087 +0000 UTC m=+3958.595615907" lastFinishedPulling="2026-02-20 09:11:52.030139403 +0000 UTC m=+3961.004634263" observedRunningTime="2026-02-20 09:11:52.674009473 +0000 UTC m=+3961.648504313" watchObservedRunningTime="2026-02-20 09:11:52.682959414 +0000 UTC m=+3961.657454234" Feb 20 09:11:58 crc kubenswrapper[4948]: I0220 09:11:58.366103 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7sjmh" Feb 20 09:11:58 crc kubenswrapper[4948]: I0220 09:11:58.366682 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7sjmh" Feb 20 09:11:58 crc kubenswrapper[4948]: I0220 09:11:58.406233 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7sjmh" Feb 20 09:11:58 crc kubenswrapper[4948]: I0220 09:11:58.757965 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7sjmh" Feb 20 09:11:58 crc kubenswrapper[4948]: I0220 09:11:58.817171 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7sjmh"] Feb 20 09:12:00 crc kubenswrapper[4948]: I0220 09:12:00.725097 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7sjmh" podUID="654b30bf-3b65-4501-9a0e-9ebc33812c73" containerName="registry-server" containerID="cri-o://e00180d9024f5ed42ba57eca358c05e3488371a087ae601107ee28e8c4a0c9d7" gracePeriod=2 Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.297466 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7sjmh" Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.420015 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4v6l\" (UniqueName: \"kubernetes.io/projected/654b30bf-3b65-4501-9a0e-9ebc33812c73-kube-api-access-g4v6l\") pod \"654b30bf-3b65-4501-9a0e-9ebc33812c73\" (UID: \"654b30bf-3b65-4501-9a0e-9ebc33812c73\") " Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.420121 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/654b30bf-3b65-4501-9a0e-9ebc33812c73-catalog-content\") pod \"654b30bf-3b65-4501-9a0e-9ebc33812c73\" (UID: \"654b30bf-3b65-4501-9a0e-9ebc33812c73\") " Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.420228 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/654b30bf-3b65-4501-9a0e-9ebc33812c73-utilities\") pod \"654b30bf-3b65-4501-9a0e-9ebc33812c73\" (UID: \"654b30bf-3b65-4501-9a0e-9ebc33812c73\") " Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.421602 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/654b30bf-3b65-4501-9a0e-9ebc33812c73-utilities" (OuterVolumeSpecName: "utilities") pod "654b30bf-3b65-4501-9a0e-9ebc33812c73" (UID: "654b30bf-3b65-4501-9a0e-9ebc33812c73"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.426466 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/654b30bf-3b65-4501-9a0e-9ebc33812c73-kube-api-access-g4v6l" (OuterVolumeSpecName: "kube-api-access-g4v6l") pod "654b30bf-3b65-4501-9a0e-9ebc33812c73" (UID: "654b30bf-3b65-4501-9a0e-9ebc33812c73"). InnerVolumeSpecName "kube-api-access-g4v6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.448434 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/654b30bf-3b65-4501-9a0e-9ebc33812c73-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "654b30bf-3b65-4501-9a0e-9ebc33812c73" (UID: "654b30bf-3b65-4501-9a0e-9ebc33812c73"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.522805 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/654b30bf-3b65-4501-9a0e-9ebc33812c73-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.523087 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/654b30bf-3b65-4501-9a0e-9ebc33812c73-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.523150 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4v6l\" (UniqueName: \"kubernetes.io/projected/654b30bf-3b65-4501-9a0e-9ebc33812c73-kube-api-access-g4v6l\") on node \"crc\" DevicePath \"\"" Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.739926 4948 generic.go:334] "Generic (PLEG): container finished" podID="654b30bf-3b65-4501-9a0e-9ebc33812c73" containerID="e00180d9024f5ed42ba57eca358c05e3488371a087ae601107ee28e8c4a0c9d7" exitCode=0 Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.740028 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sjmh" event={"ID":"654b30bf-3b65-4501-9a0e-9ebc33812c73","Type":"ContainerDied","Data":"e00180d9024f5ed42ba57eca358c05e3488371a087ae601107ee28e8c4a0c9d7"} Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.740091 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sjmh" event={"ID":"654b30bf-3b65-4501-9a0e-9ebc33812c73","Type":"ContainerDied","Data":"935397e9a001171af02db7a3be1a28e8bc26a789185e6481f73a103b9280bae0"} Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.740131 4948 scope.go:117] "RemoveContainer" containerID="e00180d9024f5ed42ba57eca358c05e3488371a087ae601107ee28e8c4a0c9d7" Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.740143 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7sjmh" Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.760615 4948 scope.go:117] "RemoveContainer" containerID="f78e60617dfa8cf4a04999e07015aa7e834279c7e0b99a1ff6c2ec26cb91d973" Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.788160 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7sjmh"] Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.790367 4948 scope.go:117] "RemoveContainer" containerID="0498136ca7009d716fd78a1af98d3ae2091af6957f0aa8b1253de62c66c87915" Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.795329 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7sjmh"] Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.822881 4948 scope.go:117] "RemoveContainer" containerID="e00180d9024f5ed42ba57eca358c05e3488371a087ae601107ee28e8c4a0c9d7" Feb 20 09:12:01 crc kubenswrapper[4948]: E0220 09:12:01.823343 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e00180d9024f5ed42ba57eca358c05e3488371a087ae601107ee28e8c4a0c9d7\": container with ID starting with e00180d9024f5ed42ba57eca358c05e3488371a087ae601107ee28e8c4a0c9d7 not found: ID does not exist" containerID="e00180d9024f5ed42ba57eca358c05e3488371a087ae601107ee28e8c4a0c9d7" Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.823378 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e00180d9024f5ed42ba57eca358c05e3488371a087ae601107ee28e8c4a0c9d7"} err="failed to get container status \"e00180d9024f5ed42ba57eca358c05e3488371a087ae601107ee28e8c4a0c9d7\": rpc error: code = NotFound desc = could not find container \"e00180d9024f5ed42ba57eca358c05e3488371a087ae601107ee28e8c4a0c9d7\": container with ID starting with e00180d9024f5ed42ba57eca358c05e3488371a087ae601107ee28e8c4a0c9d7 not found: ID does not exist" Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.823398 4948 scope.go:117] "RemoveContainer" containerID="f78e60617dfa8cf4a04999e07015aa7e834279c7e0b99a1ff6c2ec26cb91d973" Feb 20 09:12:01 crc kubenswrapper[4948]: E0220 09:12:01.823681 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f78e60617dfa8cf4a04999e07015aa7e834279c7e0b99a1ff6c2ec26cb91d973\": container with ID starting with f78e60617dfa8cf4a04999e07015aa7e834279c7e0b99a1ff6c2ec26cb91d973 not found: ID does not exist" containerID="f78e60617dfa8cf4a04999e07015aa7e834279c7e0b99a1ff6c2ec26cb91d973" Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.823711 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f78e60617dfa8cf4a04999e07015aa7e834279c7e0b99a1ff6c2ec26cb91d973"} err="failed to get container status \"f78e60617dfa8cf4a04999e07015aa7e834279c7e0b99a1ff6c2ec26cb91d973\": rpc error: code = NotFound desc = could not find container \"f78e60617dfa8cf4a04999e07015aa7e834279c7e0b99a1ff6c2ec26cb91d973\": container with ID starting with f78e60617dfa8cf4a04999e07015aa7e834279c7e0b99a1ff6c2ec26cb91d973 not found: ID does not exist" Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.823732 4948 scope.go:117] "RemoveContainer" containerID="0498136ca7009d716fd78a1af98d3ae2091af6957f0aa8b1253de62c66c87915" Feb 20 09:12:01 crc kubenswrapper[4948]: E0220 09:12:01.824013 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0498136ca7009d716fd78a1af98d3ae2091af6957f0aa8b1253de62c66c87915\": container with ID starting with 0498136ca7009d716fd78a1af98d3ae2091af6957f0aa8b1253de62c66c87915 not found: ID does not exist" containerID="0498136ca7009d716fd78a1af98d3ae2091af6957f0aa8b1253de62c66c87915" Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.824038 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0498136ca7009d716fd78a1af98d3ae2091af6957f0aa8b1253de62c66c87915"} err="failed to get container status \"0498136ca7009d716fd78a1af98d3ae2091af6957f0aa8b1253de62c66c87915\": rpc error: code = NotFound desc = could not find container \"0498136ca7009d716fd78a1af98d3ae2091af6957f0aa8b1253de62c66c87915\": container with ID starting with 0498136ca7009d716fd78a1af98d3ae2091af6957f0aa8b1253de62c66c87915 not found: ID does not exist" Feb 20 09:12:03 crc kubenswrapper[4948]: I0220 09:12:03.732082 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="654b30bf-3b65-4501-9a0e-9ebc33812c73" path="/var/lib/kubelet/pods/654b30bf-3b65-4501-9a0e-9ebc33812c73/volumes" Feb 20 09:12:16 crc kubenswrapper[4948]: E0220 09:12:16.599168 4948 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.113:40572->38.102.83.113:38489: write tcp 38.102.83.113:40572->38.102.83.113:38489: write: connection reset by peer Feb 20 09:12:38 crc kubenswrapper[4948]: I0220 09:12:38.025188 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:12:38 crc kubenswrapper[4948]: I0220 09:12:38.025698 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:13:08 crc kubenswrapper[4948]: I0220 09:13:08.024993 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:13:08 crc kubenswrapper[4948]: I0220 09:13:08.025740 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:13:31 crc kubenswrapper[4948]: I0220 09:13:31.964009 4948 generic.go:334] "Generic (PLEG): container finished" podID="d81b349b-2494-4abc-b349-44fecd284709" containerID="0b5814bb586cbca972e209566bf544539a27cba45472fac01d96fe47fe540cff" exitCode=0 Feb 20 09:13:31 crc kubenswrapper[4948]: I0220 09:13:31.964089 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ql8t9/must-gather-hvzrp" event={"ID":"d81b349b-2494-4abc-b349-44fecd284709","Type":"ContainerDied","Data":"0b5814bb586cbca972e209566bf544539a27cba45472fac01d96fe47fe540cff"} Feb 20 09:13:31 crc kubenswrapper[4948]: I0220 09:13:31.965381 4948 scope.go:117] "RemoveContainer" containerID="0b5814bb586cbca972e209566bf544539a27cba45472fac01d96fe47fe540cff" Feb 20 09:13:32 crc kubenswrapper[4948]: I0220 09:13:32.973328 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ql8t9_must-gather-hvzrp_d81b349b-2494-4abc-b349-44fecd284709/gather/0.log" Feb 20 09:13:38 crc kubenswrapper[4948]: I0220 09:13:38.025180 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:13:38 crc kubenswrapper[4948]: I0220 09:13:38.025821 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:13:38 crc kubenswrapper[4948]: I0220 09:13:38.025919 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 09:13:38 crc kubenswrapper[4948]: I0220 09:13:38.028879 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5bbfc48505968074b16a56aea08c035f5c0687419ea24b66d69a44ebd2ab9d7d"} pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 09:13:38 crc kubenswrapper[4948]: I0220 09:13:38.028996 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" containerID="cri-o://5bbfc48505968074b16a56aea08c035f5c0687419ea24b66d69a44ebd2ab9d7d" gracePeriod=600 Feb 20 09:13:39 crc kubenswrapper[4948]: I0220 09:13:39.034930 4948 generic.go:334] "Generic (PLEG): container finished" podID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerID="5bbfc48505968074b16a56aea08c035f5c0687419ea24b66d69a44ebd2ab9d7d" exitCode=0 Feb 20 09:13:39 crc kubenswrapper[4948]: I0220 09:13:39.034989 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerDied","Data":"5bbfc48505968074b16a56aea08c035f5c0687419ea24b66d69a44ebd2ab9d7d"} Feb 20 09:13:39 crc kubenswrapper[4948]: I0220 09:13:39.035498 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"2e609219eda64b1acca0e9ec84aa03d7c21986cd69e1f5ce367d12346cb500e5"} Feb 20 09:13:39 crc kubenswrapper[4948]: I0220 09:13:39.035523 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:13:45 crc kubenswrapper[4948]: I0220 09:13:45.039438 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ql8t9/must-gather-hvzrp"] Feb 20 09:13:45 crc kubenswrapper[4948]: I0220 09:13:45.040556 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-ql8t9/must-gather-hvzrp" podUID="d81b349b-2494-4abc-b349-44fecd284709" containerName="copy" containerID="cri-o://d49cb379a847549834cd028d6c373359633cf1e9d6c301526cf59089fd2bfebf" gracePeriod=2 Feb 20 09:13:45 crc kubenswrapper[4948]: I0220 09:13:45.047125 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ql8t9/must-gather-hvzrp"] Feb 20 09:13:45 crc kubenswrapper[4948]: I0220 09:13:45.489147 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ql8t9_must-gather-hvzrp_d81b349b-2494-4abc-b349-44fecd284709/copy/0.log" Feb 20 09:13:45 crc kubenswrapper[4948]: I0220 09:13:45.490107 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ql8t9/must-gather-hvzrp" Feb 20 09:13:45 crc kubenswrapper[4948]: I0220 09:13:45.602837 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d81b349b-2494-4abc-b349-44fecd284709-must-gather-output\") pod \"d81b349b-2494-4abc-b349-44fecd284709\" (UID: \"d81b349b-2494-4abc-b349-44fecd284709\") " Feb 20 09:13:45 crc kubenswrapper[4948]: I0220 09:13:45.603091 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2bcm\" (UniqueName: \"kubernetes.io/projected/d81b349b-2494-4abc-b349-44fecd284709-kube-api-access-v2bcm\") pod \"d81b349b-2494-4abc-b349-44fecd284709\" (UID: \"d81b349b-2494-4abc-b349-44fecd284709\") " Feb 20 09:13:45 crc kubenswrapper[4948]: I0220 09:13:45.610136 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d81b349b-2494-4abc-b349-44fecd284709-kube-api-access-v2bcm" (OuterVolumeSpecName: "kube-api-access-v2bcm") pod "d81b349b-2494-4abc-b349-44fecd284709" (UID: "d81b349b-2494-4abc-b349-44fecd284709"). InnerVolumeSpecName "kube-api-access-v2bcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:13:45 crc kubenswrapper[4948]: I0220 09:13:45.704765 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2bcm\" (UniqueName: \"kubernetes.io/projected/d81b349b-2494-4abc-b349-44fecd284709-kube-api-access-v2bcm\") on node \"crc\" DevicePath \"\"" Feb 20 09:13:45 crc kubenswrapper[4948]: I0220 09:13:45.770661 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d81b349b-2494-4abc-b349-44fecd284709-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "d81b349b-2494-4abc-b349-44fecd284709" (UID: "d81b349b-2494-4abc-b349-44fecd284709"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:13:45 crc kubenswrapper[4948]: I0220 09:13:45.806896 4948 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d81b349b-2494-4abc-b349-44fecd284709-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 20 09:13:46 crc kubenswrapper[4948]: I0220 09:13:46.166722 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ql8t9_must-gather-hvzrp_d81b349b-2494-4abc-b349-44fecd284709/copy/0.log" Feb 20 09:13:46 crc kubenswrapper[4948]: I0220 09:13:46.167513 4948 generic.go:334] "Generic (PLEG): container finished" podID="d81b349b-2494-4abc-b349-44fecd284709" containerID="d49cb379a847549834cd028d6c373359633cf1e9d6c301526cf59089fd2bfebf" exitCode=143 Feb 20 09:13:46 crc kubenswrapper[4948]: I0220 09:13:46.167585 4948 scope.go:117] "RemoveContainer" containerID="d49cb379a847549834cd028d6c373359633cf1e9d6c301526cf59089fd2bfebf" Feb 20 09:13:46 crc kubenswrapper[4948]: I0220 09:13:46.167725 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ql8t9/must-gather-hvzrp" Feb 20 09:13:46 crc kubenswrapper[4948]: I0220 09:13:46.191045 4948 scope.go:117] "RemoveContainer" containerID="0b5814bb586cbca972e209566bf544539a27cba45472fac01d96fe47fe540cff" Feb 20 09:13:46 crc kubenswrapper[4948]: I0220 09:13:46.287325 4948 scope.go:117] "RemoveContainer" containerID="d49cb379a847549834cd028d6c373359633cf1e9d6c301526cf59089fd2bfebf" Feb 20 09:13:46 crc kubenswrapper[4948]: E0220 09:13:46.287795 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d49cb379a847549834cd028d6c373359633cf1e9d6c301526cf59089fd2bfebf\": container with ID starting with d49cb379a847549834cd028d6c373359633cf1e9d6c301526cf59089fd2bfebf not found: ID does not exist" containerID="d49cb379a847549834cd028d6c373359633cf1e9d6c301526cf59089fd2bfebf" Feb 20 09:13:46 crc kubenswrapper[4948]: I0220 09:13:46.287823 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d49cb379a847549834cd028d6c373359633cf1e9d6c301526cf59089fd2bfebf"} err="failed to get container status \"d49cb379a847549834cd028d6c373359633cf1e9d6c301526cf59089fd2bfebf\": rpc error: code = NotFound desc = could not find container \"d49cb379a847549834cd028d6c373359633cf1e9d6c301526cf59089fd2bfebf\": container with ID starting with d49cb379a847549834cd028d6c373359633cf1e9d6c301526cf59089fd2bfebf not found: ID does not exist" Feb 20 09:13:46 crc kubenswrapper[4948]: I0220 09:13:46.287848 4948 scope.go:117] "RemoveContainer" containerID="0b5814bb586cbca972e209566bf544539a27cba45472fac01d96fe47fe540cff" Feb 20 09:13:46 crc kubenswrapper[4948]: E0220 09:13:46.288123 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b5814bb586cbca972e209566bf544539a27cba45472fac01d96fe47fe540cff\": container with ID starting with 0b5814bb586cbca972e209566bf544539a27cba45472fac01d96fe47fe540cff not found: ID does not exist" containerID="0b5814bb586cbca972e209566bf544539a27cba45472fac01d96fe47fe540cff" Feb 20 09:13:46 crc kubenswrapper[4948]: I0220 09:13:46.288147 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b5814bb586cbca972e209566bf544539a27cba45472fac01d96fe47fe540cff"} err="failed to get container status \"0b5814bb586cbca972e209566bf544539a27cba45472fac01d96fe47fe540cff\": rpc error: code = NotFound desc = could not find container \"0b5814bb586cbca972e209566bf544539a27cba45472fac01d96fe47fe540cff\": container with ID starting with 0b5814bb586cbca972e209566bf544539a27cba45472fac01d96fe47fe540cff not found: ID does not exist" Feb 20 09:13:47 crc kubenswrapper[4948]: I0220 09:13:47.745410 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d81b349b-2494-4abc-b349-44fecd284709" path="/var/lib/kubelet/pods/d81b349b-2494-4abc-b349-44fecd284709/volumes" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.207739 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c"] Feb 20 09:15:00 crc kubenswrapper[4948]: E0220 09:15:00.208803 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="654b30bf-3b65-4501-9a0e-9ebc33812c73" containerName="extract-content" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.208821 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="654b30bf-3b65-4501-9a0e-9ebc33812c73" containerName="extract-content" Feb 20 09:15:00 crc kubenswrapper[4948]: E0220 09:15:00.208834 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d81b349b-2494-4abc-b349-44fecd284709" containerName="gather" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.208843 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d81b349b-2494-4abc-b349-44fecd284709" containerName="gather" Feb 20 09:15:00 crc kubenswrapper[4948]: E0220 09:15:00.208875 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="654b30bf-3b65-4501-9a0e-9ebc33812c73" containerName="registry-server" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.208889 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="654b30bf-3b65-4501-9a0e-9ebc33812c73" containerName="registry-server" Feb 20 09:15:00 crc kubenswrapper[4948]: E0220 09:15:00.208905 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d81b349b-2494-4abc-b349-44fecd284709" containerName="copy" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.208914 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d81b349b-2494-4abc-b349-44fecd284709" containerName="copy" Feb 20 09:15:00 crc kubenswrapper[4948]: E0220 09:15:00.208946 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="654b30bf-3b65-4501-9a0e-9ebc33812c73" containerName="extract-utilities" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.208958 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="654b30bf-3b65-4501-9a0e-9ebc33812c73" containerName="extract-utilities" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.209220 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="654b30bf-3b65-4501-9a0e-9ebc33812c73" containerName="registry-server" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.209250 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d81b349b-2494-4abc-b349-44fecd284709" containerName="copy" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.209276 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d81b349b-2494-4abc-b349-44fecd284709" containerName="gather" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.210150 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.212657 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.212778 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.231541 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c"] Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.358952 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wb9ml\" (UniqueName: \"kubernetes.io/projected/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-kube-api-access-wb9ml\") pod \"collect-profiles-29526315-h789c\" (UID: \"6b98fa20-139d-4a56-b4e8-d32209bf4fb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.359213 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-secret-volume\") pod \"collect-profiles-29526315-h789c\" (UID: \"6b98fa20-139d-4a56-b4e8-d32209bf4fb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.359589 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-config-volume\") pod \"collect-profiles-29526315-h789c\" (UID: \"6b98fa20-139d-4a56-b4e8-d32209bf4fb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.461541 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wb9ml\" (UniqueName: \"kubernetes.io/projected/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-kube-api-access-wb9ml\") pod \"collect-profiles-29526315-h789c\" (UID: \"6b98fa20-139d-4a56-b4e8-d32209bf4fb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.462010 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-secret-volume\") pod \"collect-profiles-29526315-h789c\" (UID: \"6b98fa20-139d-4a56-b4e8-d32209bf4fb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.462257 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-config-volume\") pod \"collect-profiles-29526315-h789c\" (UID: \"6b98fa20-139d-4a56-b4e8-d32209bf4fb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.463110 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-config-volume\") pod \"collect-profiles-29526315-h789c\" (UID: \"6b98fa20-139d-4a56-b4e8-d32209bf4fb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.470541 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-secret-volume\") pod \"collect-profiles-29526315-h789c\" (UID: \"6b98fa20-139d-4a56-b4e8-d32209bf4fb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.479157 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wb9ml\" (UniqueName: \"kubernetes.io/projected/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-kube-api-access-wb9ml\") pod \"collect-profiles-29526315-h789c\" (UID: \"6b98fa20-139d-4a56-b4e8-d32209bf4fb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.537768 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c" Feb 20 09:15:01 crc kubenswrapper[4948]: I0220 09:15:01.033829 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c"] Feb 20 09:15:01 crc kubenswrapper[4948]: I0220 09:15:01.937212 4948 generic.go:334] "Generic (PLEG): container finished" podID="6b98fa20-139d-4a56-b4e8-d32209bf4fb4" containerID="a70d23d0cc71f3b65f68123bbc28e9d24e5f9ab9fb49d3e0ece37db0403b1cdc" exitCode=0 Feb 20 09:15:01 crc kubenswrapper[4948]: I0220 09:15:01.937268 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c" event={"ID":"6b98fa20-139d-4a56-b4e8-d32209bf4fb4","Type":"ContainerDied","Data":"a70d23d0cc71f3b65f68123bbc28e9d24e5f9ab9fb49d3e0ece37db0403b1cdc"} Feb 20 09:15:01 crc kubenswrapper[4948]: I0220 09:15:01.937501 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c" event={"ID":"6b98fa20-139d-4a56-b4e8-d32209bf4fb4","Type":"ContainerStarted","Data":"03e4e4d97c682186adfd9949701d3c54e01060de77e731ddb5f428f35ac442a8"} Feb 20 09:15:03 crc kubenswrapper[4948]: I0220 09:15:03.328012 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c" Feb 20 09:15:03 crc kubenswrapper[4948]: I0220 09:15:03.421948 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-config-volume\") pod \"6b98fa20-139d-4a56-b4e8-d32209bf4fb4\" (UID: \"6b98fa20-139d-4a56-b4e8-d32209bf4fb4\") " Feb 20 09:15:03 crc kubenswrapper[4948]: I0220 09:15:03.422169 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wb9ml\" (UniqueName: \"kubernetes.io/projected/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-kube-api-access-wb9ml\") pod \"6b98fa20-139d-4a56-b4e8-d32209bf4fb4\" (UID: \"6b98fa20-139d-4a56-b4e8-d32209bf4fb4\") " Feb 20 09:15:03 crc kubenswrapper[4948]: I0220 09:15:03.422206 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-secret-volume\") pod \"6b98fa20-139d-4a56-b4e8-d32209bf4fb4\" (UID: \"6b98fa20-139d-4a56-b4e8-d32209bf4fb4\") " Feb 20 09:15:03 crc kubenswrapper[4948]: I0220 09:15:03.422808 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-config-volume" (OuterVolumeSpecName: "config-volume") pod "6b98fa20-139d-4a56-b4e8-d32209bf4fb4" (UID: "6b98fa20-139d-4a56-b4e8-d32209bf4fb4"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 09:15:03 crc kubenswrapper[4948]: I0220 09:15:03.429047 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6b98fa20-139d-4a56-b4e8-d32209bf4fb4" (UID: "6b98fa20-139d-4a56-b4e8-d32209bf4fb4"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 09:15:03 crc kubenswrapper[4948]: I0220 09:15:03.429284 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-kube-api-access-wb9ml" (OuterVolumeSpecName: "kube-api-access-wb9ml") pod "6b98fa20-139d-4a56-b4e8-d32209bf4fb4" (UID: "6b98fa20-139d-4a56-b4e8-d32209bf4fb4"). InnerVolumeSpecName "kube-api-access-wb9ml". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:15:03 crc kubenswrapper[4948]: I0220 09:15:03.524723 4948 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-config-volume\") on node \"crc\" DevicePath \"\"" Feb 20 09:15:03 crc kubenswrapper[4948]: I0220 09:15:03.524997 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wb9ml\" (UniqueName: \"kubernetes.io/projected/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-kube-api-access-wb9ml\") on node \"crc\" DevicePath \"\"" Feb 20 09:15:03 crc kubenswrapper[4948]: I0220 09:15:03.525008 4948 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 20 09:15:03 crc kubenswrapper[4948]: I0220 09:15:03.971661 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c" event={"ID":"6b98fa20-139d-4a56-b4e8-d32209bf4fb4","Type":"ContainerDied","Data":"03e4e4d97c682186adfd9949701d3c54e01060de77e731ddb5f428f35ac442a8"} Feb 20 09:15:03 crc kubenswrapper[4948]: I0220 09:15:03.971720 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03e4e4d97c682186adfd9949701d3c54e01060de77e731ddb5f428f35ac442a8" Feb 20 09:15:03 crc kubenswrapper[4948]: I0220 09:15:03.971743 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c" Feb 20 09:15:04 crc kubenswrapper[4948]: I0220 09:15:04.421054 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm"] Feb 20 09:15:04 crc kubenswrapper[4948]: I0220 09:15:04.431928 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm"] Feb 20 09:15:05 crc kubenswrapper[4948]: I0220 09:15:05.738487 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2aaa3d0b-4371-404a-8d94-9a7158fff416" path="/var/lib/kubelet/pods/2aaa3d0b-4371-404a-8d94-9a7158fff416/volumes" Feb 20 09:15:15 crc kubenswrapper[4948]: I0220 09:15:15.377131 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9mdkj"] Feb 20 09:15:15 crc kubenswrapper[4948]: E0220 09:15:15.378644 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b98fa20-139d-4a56-b4e8-d32209bf4fb4" containerName="collect-profiles" Feb 20 09:15:15 crc kubenswrapper[4948]: I0220 09:15:15.378671 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b98fa20-139d-4a56-b4e8-d32209bf4fb4" containerName="collect-profiles" Feb 20 09:15:15 crc kubenswrapper[4948]: I0220 09:15:15.379221 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b98fa20-139d-4a56-b4e8-d32209bf4fb4" containerName="collect-profiles" Feb 20 09:15:15 crc kubenswrapper[4948]: I0220 09:15:15.382246 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9mdkj" Feb 20 09:15:15 crc kubenswrapper[4948]: I0220 09:15:15.404020 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9mdkj"] Feb 20 09:15:15 crc kubenswrapper[4948]: I0220 09:15:15.499132 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1443dd18-dcab-42e8-8592-63e1ae18a6ef-catalog-content\") pod \"certified-operators-9mdkj\" (UID: \"1443dd18-dcab-42e8-8592-63e1ae18a6ef\") " pod="openshift-marketplace/certified-operators-9mdkj" Feb 20 09:15:15 crc kubenswrapper[4948]: I0220 09:15:15.499278 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1443dd18-dcab-42e8-8592-63e1ae18a6ef-utilities\") pod \"certified-operators-9mdkj\" (UID: \"1443dd18-dcab-42e8-8592-63e1ae18a6ef\") " pod="openshift-marketplace/certified-operators-9mdkj" Feb 20 09:15:15 crc kubenswrapper[4948]: I0220 09:15:15.499603 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwrms\" (UniqueName: \"kubernetes.io/projected/1443dd18-dcab-42e8-8592-63e1ae18a6ef-kube-api-access-xwrms\") pod \"certified-operators-9mdkj\" (UID: \"1443dd18-dcab-42e8-8592-63e1ae18a6ef\") " pod="openshift-marketplace/certified-operators-9mdkj" Feb 20 09:15:15 crc kubenswrapper[4948]: I0220 09:15:15.601802 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1443dd18-dcab-42e8-8592-63e1ae18a6ef-catalog-content\") pod \"certified-operators-9mdkj\" (UID: \"1443dd18-dcab-42e8-8592-63e1ae18a6ef\") " pod="openshift-marketplace/certified-operators-9mdkj" Feb 20 09:15:15 crc kubenswrapper[4948]: I0220 09:15:15.601883 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1443dd18-dcab-42e8-8592-63e1ae18a6ef-utilities\") pod \"certified-operators-9mdkj\" (UID: \"1443dd18-dcab-42e8-8592-63e1ae18a6ef\") " pod="openshift-marketplace/certified-operators-9mdkj" Feb 20 09:15:15 crc kubenswrapper[4948]: I0220 09:15:15.602027 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwrms\" (UniqueName: \"kubernetes.io/projected/1443dd18-dcab-42e8-8592-63e1ae18a6ef-kube-api-access-xwrms\") pod \"certified-operators-9mdkj\" (UID: \"1443dd18-dcab-42e8-8592-63e1ae18a6ef\") " pod="openshift-marketplace/certified-operators-9mdkj" Feb 20 09:15:15 crc kubenswrapper[4948]: I0220 09:15:15.602552 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1443dd18-dcab-42e8-8592-63e1ae18a6ef-catalog-content\") pod \"certified-operators-9mdkj\" (UID: \"1443dd18-dcab-42e8-8592-63e1ae18a6ef\") " pod="openshift-marketplace/certified-operators-9mdkj" Feb 20 09:15:15 crc kubenswrapper[4948]: I0220 09:15:15.602732 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1443dd18-dcab-42e8-8592-63e1ae18a6ef-utilities\") pod \"certified-operators-9mdkj\" (UID: \"1443dd18-dcab-42e8-8592-63e1ae18a6ef\") " pod="openshift-marketplace/certified-operators-9mdkj" Feb 20 09:15:15 crc kubenswrapper[4948]: I0220 09:15:15.626694 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwrms\" (UniqueName: \"kubernetes.io/projected/1443dd18-dcab-42e8-8592-63e1ae18a6ef-kube-api-access-xwrms\") pod \"certified-operators-9mdkj\" (UID: \"1443dd18-dcab-42e8-8592-63e1ae18a6ef\") " pod="openshift-marketplace/certified-operators-9mdkj" Feb 20 09:15:15 crc kubenswrapper[4948]: I0220 09:15:15.711210 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9mdkj" Feb 20 09:15:16 crc kubenswrapper[4948]: I0220 09:15:16.222299 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9mdkj"] Feb 20 09:15:17 crc kubenswrapper[4948]: I0220 09:15:17.142434 4948 generic.go:334] "Generic (PLEG): container finished" podID="1443dd18-dcab-42e8-8592-63e1ae18a6ef" containerID="6cacdffb656dd31edb857180139ac055fde2c217ffc1c87e516b8f4b9092b59c" exitCode=0 Feb 20 09:15:17 crc kubenswrapper[4948]: I0220 09:15:17.142500 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9mdkj" event={"ID":"1443dd18-dcab-42e8-8592-63e1ae18a6ef","Type":"ContainerDied","Data":"6cacdffb656dd31edb857180139ac055fde2c217ffc1c87e516b8f4b9092b59c"} Feb 20 09:15:17 crc kubenswrapper[4948]: I0220 09:15:17.142557 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9mdkj" event={"ID":"1443dd18-dcab-42e8-8592-63e1ae18a6ef","Type":"ContainerStarted","Data":"f5de5732c2b7595df1124e25393432adf4d492acc7ff2a3bfd2cf5030afc2d33"} Feb 20 09:15:18 crc kubenswrapper[4948]: I0220 09:15:18.158143 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9mdkj" event={"ID":"1443dd18-dcab-42e8-8592-63e1ae18a6ef","Type":"ContainerStarted","Data":"e4d2e8ef702be4e555465ec7a1aaea879fb47d956b0f098c18df270181ca9042"} Feb 20 09:15:19 crc kubenswrapper[4948]: I0220 09:15:19.171320 4948 generic.go:334] "Generic (PLEG): container finished" podID="1443dd18-dcab-42e8-8592-63e1ae18a6ef" containerID="e4d2e8ef702be4e555465ec7a1aaea879fb47d956b0f098c18df270181ca9042" exitCode=0 Feb 20 09:15:19 crc kubenswrapper[4948]: I0220 09:15:19.171402 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9mdkj" event={"ID":"1443dd18-dcab-42e8-8592-63e1ae18a6ef","Type":"ContainerDied","Data":"e4d2e8ef702be4e555465ec7a1aaea879fb47d956b0f098c18df270181ca9042"} Feb 20 09:15:20 crc kubenswrapper[4948]: I0220 09:15:20.186127 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9mdkj" event={"ID":"1443dd18-dcab-42e8-8592-63e1ae18a6ef","Type":"ContainerStarted","Data":"011cc2bc06960fc14c7f8fb4e578e2b7d990552c5f3d14d6bd6095826445bf0a"} Feb 20 09:15:20 crc kubenswrapper[4948]: I0220 09:15:20.208519 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9mdkj" podStartSLOduration=2.743889042 podStartE2EDuration="5.20849947s" podCreationTimestamp="2026-02-20 09:15:15 +0000 UTC" firstStartedPulling="2026-02-20 09:15:17.145506114 +0000 UTC m=+4166.120000974" lastFinishedPulling="2026-02-20 09:15:19.610116542 +0000 UTC m=+4168.584611402" observedRunningTime="2026-02-20 09:15:20.206228754 +0000 UTC m=+4169.180723614" watchObservedRunningTime="2026-02-20 09:15:20.20849947 +0000 UTC m=+4169.182994290" Feb 20 09:15:20 crc kubenswrapper[4948]: I0220 09:15:20.349311 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-d28wc"] Feb 20 09:15:20 crc kubenswrapper[4948]: I0220 09:15:20.353300 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d28wc" Feb 20 09:15:20 crc kubenswrapper[4948]: I0220 09:15:20.361643 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d28wc"] Feb 20 09:15:20 crc kubenswrapper[4948]: I0220 09:15:20.513305 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4425c869-8c06-4583-bd41-9fbb012a893e-utilities\") pod \"community-operators-d28wc\" (UID: \"4425c869-8c06-4583-bd41-9fbb012a893e\") " pod="openshift-marketplace/community-operators-d28wc" Feb 20 09:15:20 crc kubenswrapper[4948]: I0220 09:15:20.513471 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rl724\" (UniqueName: \"kubernetes.io/projected/4425c869-8c06-4583-bd41-9fbb012a893e-kube-api-access-rl724\") pod \"community-operators-d28wc\" (UID: \"4425c869-8c06-4583-bd41-9fbb012a893e\") " pod="openshift-marketplace/community-operators-d28wc" Feb 20 09:15:20 crc kubenswrapper[4948]: I0220 09:15:20.513508 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4425c869-8c06-4583-bd41-9fbb012a893e-catalog-content\") pod \"community-operators-d28wc\" (UID: \"4425c869-8c06-4583-bd41-9fbb012a893e\") " pod="openshift-marketplace/community-operators-d28wc" Feb 20 09:15:20 crc kubenswrapper[4948]: I0220 09:15:20.614840 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rl724\" (UniqueName: \"kubernetes.io/projected/4425c869-8c06-4583-bd41-9fbb012a893e-kube-api-access-rl724\") pod \"community-operators-d28wc\" (UID: \"4425c869-8c06-4583-bd41-9fbb012a893e\") " pod="openshift-marketplace/community-operators-d28wc" Feb 20 09:15:20 crc kubenswrapper[4948]: I0220 09:15:20.614920 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4425c869-8c06-4583-bd41-9fbb012a893e-catalog-content\") pod \"community-operators-d28wc\" (UID: \"4425c869-8c06-4583-bd41-9fbb012a893e\") " pod="openshift-marketplace/community-operators-d28wc" Feb 20 09:15:20 crc kubenswrapper[4948]: I0220 09:15:20.615113 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4425c869-8c06-4583-bd41-9fbb012a893e-utilities\") pod \"community-operators-d28wc\" (UID: \"4425c869-8c06-4583-bd41-9fbb012a893e\") " pod="openshift-marketplace/community-operators-d28wc" Feb 20 09:15:20 crc kubenswrapper[4948]: I0220 09:15:20.615599 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4425c869-8c06-4583-bd41-9fbb012a893e-utilities\") pod \"community-operators-d28wc\" (UID: \"4425c869-8c06-4583-bd41-9fbb012a893e\") " pod="openshift-marketplace/community-operators-d28wc" Feb 20 09:15:20 crc kubenswrapper[4948]: I0220 09:15:20.615695 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4425c869-8c06-4583-bd41-9fbb012a893e-catalog-content\") pod \"community-operators-d28wc\" (UID: \"4425c869-8c06-4583-bd41-9fbb012a893e\") " pod="openshift-marketplace/community-operators-d28wc" Feb 20 09:15:20 crc kubenswrapper[4948]: I0220 09:15:20.639710 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rl724\" (UniqueName: \"kubernetes.io/projected/4425c869-8c06-4583-bd41-9fbb012a893e-kube-api-access-rl724\") pod \"community-operators-d28wc\" (UID: \"4425c869-8c06-4583-bd41-9fbb012a893e\") " pod="openshift-marketplace/community-operators-d28wc" Feb 20 09:15:20 crc kubenswrapper[4948]: I0220 09:15:20.686443 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d28wc" Feb 20 09:15:21 crc kubenswrapper[4948]: I0220 09:15:21.214002 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d28wc"] Feb 20 09:15:21 crc kubenswrapper[4948]: W0220 09:15:21.222405 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4425c869_8c06_4583_bd41_9fbb012a893e.slice/crio-5b8038ad2561ab00e000949c744573df91dd4c5894d48f3525defea5f9208cb0 WatchSource:0}: Error finding container 5b8038ad2561ab00e000949c744573df91dd4c5894d48f3525defea5f9208cb0: Status 404 returned error can't find the container with id 5b8038ad2561ab00e000949c744573df91dd4c5894d48f3525defea5f9208cb0 Feb 20 09:15:22 crc kubenswrapper[4948]: I0220 09:15:22.212331 4948 generic.go:334] "Generic (PLEG): container finished" podID="4425c869-8c06-4583-bd41-9fbb012a893e" containerID="5ddb552b30bf71c7600bd2cc910fb0b0c766c877286178ccb42d4a02b77e9722" exitCode=0 Feb 20 09:15:22 crc kubenswrapper[4948]: I0220 09:15:22.213446 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d28wc" event={"ID":"4425c869-8c06-4583-bd41-9fbb012a893e","Type":"ContainerDied","Data":"5ddb552b30bf71c7600bd2cc910fb0b0c766c877286178ccb42d4a02b77e9722"} Feb 20 09:15:22 crc kubenswrapper[4948]: I0220 09:15:22.213684 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d28wc" event={"ID":"4425c869-8c06-4583-bd41-9fbb012a893e","Type":"ContainerStarted","Data":"5b8038ad2561ab00e000949c744573df91dd4c5894d48f3525defea5f9208cb0"} Feb 20 09:15:24 crc kubenswrapper[4948]: I0220 09:15:24.241469 4948 generic.go:334] "Generic (PLEG): container finished" podID="4425c869-8c06-4583-bd41-9fbb012a893e" containerID="3cc32df17d4cb5705d3995090fdf4ac05ae48ad237b349fcecac92279bfda242" exitCode=0 Feb 20 09:15:24 crc kubenswrapper[4948]: I0220 09:15:24.242018 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d28wc" event={"ID":"4425c869-8c06-4583-bd41-9fbb012a893e","Type":"ContainerDied","Data":"3cc32df17d4cb5705d3995090fdf4ac05ae48ad237b349fcecac92279bfda242"} Feb 20 09:15:25 crc kubenswrapper[4948]: I0220 09:15:25.257588 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d28wc" event={"ID":"4425c869-8c06-4583-bd41-9fbb012a893e","Type":"ContainerStarted","Data":"edffd0bfb746bfdf03464481477fac6fcf1088c83d307d3e9e259c8480f089be"} Feb 20 09:15:25 crc kubenswrapper[4948]: I0220 09:15:25.287798 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-d28wc" podStartSLOduration=2.886426821 podStartE2EDuration="5.287775047s" podCreationTimestamp="2026-02-20 09:15:20 +0000 UTC" firstStartedPulling="2026-02-20 09:15:22.224682419 +0000 UTC m=+4171.199177279" lastFinishedPulling="2026-02-20 09:15:24.626030645 +0000 UTC m=+4173.600525505" observedRunningTime="2026-02-20 09:15:25.277046352 +0000 UTC m=+4174.251541182" watchObservedRunningTime="2026-02-20 09:15:25.287775047 +0000 UTC m=+4174.262269887" Feb 20 09:15:25 crc kubenswrapper[4948]: I0220 09:15:25.712294 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9mdkj" Feb 20 09:15:25 crc kubenswrapper[4948]: I0220 09:15:25.712357 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9mdkj" Feb 20 09:15:25 crc kubenswrapper[4948]: I0220 09:15:25.774882 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9mdkj" Feb 20 09:15:26 crc kubenswrapper[4948]: I0220 09:15:26.353486 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9mdkj" Feb 20 09:15:26 crc kubenswrapper[4948]: I0220 09:15:26.930437 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9mdkj"] Feb 20 09:15:28 crc kubenswrapper[4948]: I0220 09:15:28.290290 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9mdkj" podUID="1443dd18-dcab-42e8-8592-63e1ae18a6ef" containerName="registry-server" containerID="cri-o://011cc2bc06960fc14c7f8fb4e578e2b7d990552c5f3d14d6bd6095826445bf0a" gracePeriod=2 Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.267413 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9mdkj" Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.301643 4948 generic.go:334] "Generic (PLEG): container finished" podID="1443dd18-dcab-42e8-8592-63e1ae18a6ef" containerID="011cc2bc06960fc14c7f8fb4e578e2b7d990552c5f3d14d6bd6095826445bf0a" exitCode=0 Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.301690 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9mdkj" event={"ID":"1443dd18-dcab-42e8-8592-63e1ae18a6ef","Type":"ContainerDied","Data":"011cc2bc06960fc14c7f8fb4e578e2b7d990552c5f3d14d6bd6095826445bf0a"} Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.301706 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9mdkj" Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.301734 4948 scope.go:117] "RemoveContainer" containerID="011cc2bc06960fc14c7f8fb4e578e2b7d990552c5f3d14d6bd6095826445bf0a" Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.301722 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9mdkj" event={"ID":"1443dd18-dcab-42e8-8592-63e1ae18a6ef","Type":"ContainerDied","Data":"f5de5732c2b7595df1124e25393432adf4d492acc7ff2a3bfd2cf5030afc2d33"} Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.321212 4948 scope.go:117] "RemoveContainer" containerID="e4d2e8ef702be4e555465ec7a1aaea879fb47d956b0f098c18df270181ca9042" Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.348924 4948 scope.go:117] "RemoveContainer" containerID="6cacdffb656dd31edb857180139ac055fde2c217ffc1c87e516b8f4b9092b59c" Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.386185 4948 scope.go:117] "RemoveContainer" containerID="011cc2bc06960fc14c7f8fb4e578e2b7d990552c5f3d14d6bd6095826445bf0a" Feb 20 09:15:29 crc kubenswrapper[4948]: E0220 09:15:29.386662 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"011cc2bc06960fc14c7f8fb4e578e2b7d990552c5f3d14d6bd6095826445bf0a\": container with ID starting with 011cc2bc06960fc14c7f8fb4e578e2b7d990552c5f3d14d6bd6095826445bf0a not found: ID does not exist" containerID="011cc2bc06960fc14c7f8fb4e578e2b7d990552c5f3d14d6bd6095826445bf0a" Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.386703 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"011cc2bc06960fc14c7f8fb4e578e2b7d990552c5f3d14d6bd6095826445bf0a"} err="failed to get container status \"011cc2bc06960fc14c7f8fb4e578e2b7d990552c5f3d14d6bd6095826445bf0a\": rpc error: code = NotFound desc = could not find container \"011cc2bc06960fc14c7f8fb4e578e2b7d990552c5f3d14d6bd6095826445bf0a\": container with ID starting with 011cc2bc06960fc14c7f8fb4e578e2b7d990552c5f3d14d6bd6095826445bf0a not found: ID does not exist" Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.386729 4948 scope.go:117] "RemoveContainer" containerID="e4d2e8ef702be4e555465ec7a1aaea879fb47d956b0f098c18df270181ca9042" Feb 20 09:15:29 crc kubenswrapper[4948]: E0220 09:15:29.387356 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4d2e8ef702be4e555465ec7a1aaea879fb47d956b0f098c18df270181ca9042\": container with ID starting with e4d2e8ef702be4e555465ec7a1aaea879fb47d956b0f098c18df270181ca9042 not found: ID does not exist" containerID="e4d2e8ef702be4e555465ec7a1aaea879fb47d956b0f098c18df270181ca9042" Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.387455 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4d2e8ef702be4e555465ec7a1aaea879fb47d956b0f098c18df270181ca9042"} err="failed to get container status \"e4d2e8ef702be4e555465ec7a1aaea879fb47d956b0f098c18df270181ca9042\": rpc error: code = NotFound desc = could not find container \"e4d2e8ef702be4e555465ec7a1aaea879fb47d956b0f098c18df270181ca9042\": container with ID starting with e4d2e8ef702be4e555465ec7a1aaea879fb47d956b0f098c18df270181ca9042 not found: ID does not exist" Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.387511 4948 scope.go:117] "RemoveContainer" containerID="6cacdffb656dd31edb857180139ac055fde2c217ffc1c87e516b8f4b9092b59c" Feb 20 09:15:29 crc kubenswrapper[4948]: E0220 09:15:29.387877 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6cacdffb656dd31edb857180139ac055fde2c217ffc1c87e516b8f4b9092b59c\": container with ID starting with 6cacdffb656dd31edb857180139ac055fde2c217ffc1c87e516b8f4b9092b59c not found: ID does not exist" containerID="6cacdffb656dd31edb857180139ac055fde2c217ffc1c87e516b8f4b9092b59c" Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.387943 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6cacdffb656dd31edb857180139ac055fde2c217ffc1c87e516b8f4b9092b59c"} err="failed to get container status \"6cacdffb656dd31edb857180139ac055fde2c217ffc1c87e516b8f4b9092b59c\": rpc error: code = NotFound desc = could not find container \"6cacdffb656dd31edb857180139ac055fde2c217ffc1c87e516b8f4b9092b59c\": container with ID starting with 6cacdffb656dd31edb857180139ac055fde2c217ffc1c87e516b8f4b9092b59c not found: ID does not exist" Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.406559 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1443dd18-dcab-42e8-8592-63e1ae18a6ef-utilities\") pod \"1443dd18-dcab-42e8-8592-63e1ae18a6ef\" (UID: \"1443dd18-dcab-42e8-8592-63e1ae18a6ef\") " Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.406688 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1443dd18-dcab-42e8-8592-63e1ae18a6ef-catalog-content\") pod \"1443dd18-dcab-42e8-8592-63e1ae18a6ef\" (UID: \"1443dd18-dcab-42e8-8592-63e1ae18a6ef\") " Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.406734 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwrms\" (UniqueName: \"kubernetes.io/projected/1443dd18-dcab-42e8-8592-63e1ae18a6ef-kube-api-access-xwrms\") pod \"1443dd18-dcab-42e8-8592-63e1ae18a6ef\" (UID: \"1443dd18-dcab-42e8-8592-63e1ae18a6ef\") " Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.417385 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1443dd18-dcab-42e8-8592-63e1ae18a6ef-kube-api-access-xwrms" (OuterVolumeSpecName: "kube-api-access-xwrms") pod "1443dd18-dcab-42e8-8592-63e1ae18a6ef" (UID: "1443dd18-dcab-42e8-8592-63e1ae18a6ef"). InnerVolumeSpecName "kube-api-access-xwrms". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.418538 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1443dd18-dcab-42e8-8592-63e1ae18a6ef-utilities" (OuterVolumeSpecName: "utilities") pod "1443dd18-dcab-42e8-8592-63e1ae18a6ef" (UID: "1443dd18-dcab-42e8-8592-63e1ae18a6ef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.477374 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1443dd18-dcab-42e8-8592-63e1ae18a6ef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1443dd18-dcab-42e8-8592-63e1ae18a6ef" (UID: "1443dd18-dcab-42e8-8592-63e1ae18a6ef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.508613 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1443dd18-dcab-42e8-8592-63e1ae18a6ef-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.508639 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1443dd18-dcab-42e8-8592-63e1ae18a6ef-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.508651 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwrms\" (UniqueName: \"kubernetes.io/projected/1443dd18-dcab-42e8-8592-63e1ae18a6ef-kube-api-access-xwrms\") on node \"crc\" DevicePath \"\"" Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.671176 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9mdkj"] Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.681744 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9mdkj"] Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.746363 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1443dd18-dcab-42e8-8592-63e1ae18a6ef" path="/var/lib/kubelet/pods/1443dd18-dcab-42e8-8592-63e1ae18a6ef/volumes" Feb 20 09:15:30 crc kubenswrapper[4948]: I0220 09:15:30.686546 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-d28wc" Feb 20 09:15:30 crc kubenswrapper[4948]: I0220 09:15:30.686938 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-d28wc" Feb 20 09:15:30 crc kubenswrapper[4948]: I0220 09:15:30.756412 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-d28wc" Feb 20 09:15:31 crc kubenswrapper[4948]: I0220 09:15:31.399025 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-d28wc" Feb 20 09:15:32 crc kubenswrapper[4948]: I0220 09:15:32.534835 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d28wc"] Feb 20 09:15:33 crc kubenswrapper[4948]: I0220 09:15:33.351282 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-d28wc" podUID="4425c869-8c06-4583-bd41-9fbb012a893e" containerName="registry-server" containerID="cri-o://edffd0bfb746bfdf03464481477fac6fcf1088c83d307d3e9e259c8480f089be" gracePeriod=2 Feb 20 09:15:33 crc kubenswrapper[4948]: I0220 09:15:33.866063 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d28wc" Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.007469 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rl724\" (UniqueName: \"kubernetes.io/projected/4425c869-8c06-4583-bd41-9fbb012a893e-kube-api-access-rl724\") pod \"4425c869-8c06-4583-bd41-9fbb012a893e\" (UID: \"4425c869-8c06-4583-bd41-9fbb012a893e\") " Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.007683 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4425c869-8c06-4583-bd41-9fbb012a893e-catalog-content\") pod \"4425c869-8c06-4583-bd41-9fbb012a893e\" (UID: \"4425c869-8c06-4583-bd41-9fbb012a893e\") " Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.007845 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4425c869-8c06-4583-bd41-9fbb012a893e-utilities\") pod \"4425c869-8c06-4583-bd41-9fbb012a893e\" (UID: \"4425c869-8c06-4583-bd41-9fbb012a893e\") " Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.009123 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4425c869-8c06-4583-bd41-9fbb012a893e-utilities" (OuterVolumeSpecName: "utilities") pod "4425c869-8c06-4583-bd41-9fbb012a893e" (UID: "4425c869-8c06-4583-bd41-9fbb012a893e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.013865 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4425c869-8c06-4583-bd41-9fbb012a893e-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.016354 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4425c869-8c06-4583-bd41-9fbb012a893e-kube-api-access-rl724" (OuterVolumeSpecName: "kube-api-access-rl724") pod "4425c869-8c06-4583-bd41-9fbb012a893e" (UID: "4425c869-8c06-4583-bd41-9fbb012a893e"). InnerVolumeSpecName "kube-api-access-rl724". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.117503 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rl724\" (UniqueName: \"kubernetes.io/projected/4425c869-8c06-4583-bd41-9fbb012a893e-kube-api-access-rl724\") on node \"crc\" DevicePath \"\"" Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.266878 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4425c869-8c06-4583-bd41-9fbb012a893e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4425c869-8c06-4583-bd41-9fbb012a893e" (UID: "4425c869-8c06-4583-bd41-9fbb012a893e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.322046 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4425c869-8c06-4583-bd41-9fbb012a893e-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.368300 4948 generic.go:334] "Generic (PLEG): container finished" podID="4425c869-8c06-4583-bd41-9fbb012a893e" containerID="edffd0bfb746bfdf03464481477fac6fcf1088c83d307d3e9e259c8480f089be" exitCode=0 Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.368347 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d28wc" event={"ID":"4425c869-8c06-4583-bd41-9fbb012a893e","Type":"ContainerDied","Data":"edffd0bfb746bfdf03464481477fac6fcf1088c83d307d3e9e259c8480f089be"} Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.368408 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d28wc" event={"ID":"4425c869-8c06-4583-bd41-9fbb012a893e","Type":"ContainerDied","Data":"5b8038ad2561ab00e000949c744573df91dd4c5894d48f3525defea5f9208cb0"} Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.368434 4948 scope.go:117] "RemoveContainer" containerID="edffd0bfb746bfdf03464481477fac6fcf1088c83d307d3e9e259c8480f089be" Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.368540 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d28wc" Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.405672 4948 scope.go:117] "RemoveContainer" containerID="3cc32df17d4cb5705d3995090fdf4ac05ae48ad237b349fcecac92279bfda242" Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.425834 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d28wc"] Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.439149 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-d28wc"] Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.445619 4948 scope.go:117] "RemoveContainer" containerID="5ddb552b30bf71c7600bd2cc910fb0b0c766c877286178ccb42d4a02b77e9722" Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.490791 4948 scope.go:117] "RemoveContainer" containerID="edffd0bfb746bfdf03464481477fac6fcf1088c83d307d3e9e259c8480f089be" Feb 20 09:15:34 crc kubenswrapper[4948]: E0220 09:15:34.491414 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edffd0bfb746bfdf03464481477fac6fcf1088c83d307d3e9e259c8480f089be\": container with ID starting with edffd0bfb746bfdf03464481477fac6fcf1088c83d307d3e9e259c8480f089be not found: ID does not exist" containerID="edffd0bfb746bfdf03464481477fac6fcf1088c83d307d3e9e259c8480f089be" Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.491662 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edffd0bfb746bfdf03464481477fac6fcf1088c83d307d3e9e259c8480f089be"} err="failed to get container status \"edffd0bfb746bfdf03464481477fac6fcf1088c83d307d3e9e259c8480f089be\": rpc error: code = NotFound desc = could not find container \"edffd0bfb746bfdf03464481477fac6fcf1088c83d307d3e9e259c8480f089be\": container with ID starting with edffd0bfb746bfdf03464481477fac6fcf1088c83d307d3e9e259c8480f089be not found: ID does not exist" Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.491891 4948 scope.go:117] "RemoveContainer" containerID="3cc32df17d4cb5705d3995090fdf4ac05ae48ad237b349fcecac92279bfda242" Feb 20 09:15:34 crc kubenswrapper[4948]: E0220 09:15:34.493308 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3cc32df17d4cb5705d3995090fdf4ac05ae48ad237b349fcecac92279bfda242\": container with ID starting with 3cc32df17d4cb5705d3995090fdf4ac05ae48ad237b349fcecac92279bfda242 not found: ID does not exist" containerID="3cc32df17d4cb5705d3995090fdf4ac05ae48ad237b349fcecac92279bfda242" Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.493350 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3cc32df17d4cb5705d3995090fdf4ac05ae48ad237b349fcecac92279bfda242"} err="failed to get container status \"3cc32df17d4cb5705d3995090fdf4ac05ae48ad237b349fcecac92279bfda242\": rpc error: code = NotFound desc = could not find container \"3cc32df17d4cb5705d3995090fdf4ac05ae48ad237b349fcecac92279bfda242\": container with ID starting with 3cc32df17d4cb5705d3995090fdf4ac05ae48ad237b349fcecac92279bfda242 not found: ID does not exist" Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.493399 4948 scope.go:117] "RemoveContainer" containerID="5ddb552b30bf71c7600bd2cc910fb0b0c766c877286178ccb42d4a02b77e9722" Feb 20 09:15:34 crc kubenswrapper[4948]: E0220 09:15:34.493784 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ddb552b30bf71c7600bd2cc910fb0b0c766c877286178ccb42d4a02b77e9722\": container with ID starting with 5ddb552b30bf71c7600bd2cc910fb0b0c766c877286178ccb42d4a02b77e9722 not found: ID does not exist" containerID="5ddb552b30bf71c7600bd2cc910fb0b0c766c877286178ccb42d4a02b77e9722" Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.493821 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ddb552b30bf71c7600bd2cc910fb0b0c766c877286178ccb42d4a02b77e9722"} err="failed to get container status \"5ddb552b30bf71c7600bd2cc910fb0b0c766c877286178ccb42d4a02b77e9722\": rpc error: code = NotFound desc = could not find container \"5ddb552b30bf71c7600bd2cc910fb0b0c766c877286178ccb42d4a02b77e9722\": container with ID starting with 5ddb552b30bf71c7600bd2cc910fb0b0c766c877286178ccb42d4a02b77e9722 not found: ID does not exist" Feb 20 09:15:35 crc kubenswrapper[4948]: I0220 09:15:35.736680 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4425c869-8c06-4583-bd41-9fbb012a893e" path="/var/lib/kubelet/pods/4425c869-8c06-4583-bd41-9fbb012a893e/volumes" Feb 20 09:15:38 crc kubenswrapper[4948]: I0220 09:15:38.025477 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:15:38 crc kubenswrapper[4948]: I0220 09:15:38.026316 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:16:03 crc kubenswrapper[4948]: I0220 09:16:03.417886 4948 scope.go:117] "RemoveContainer" containerID="21765550ece8828e59342442ce6d5be7d08bd1dbcd30f8abbcc286c1bfd5f713" Feb 20 09:16:08 crc kubenswrapper[4948]: I0220 09:16:08.025256 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:16:08 crc kubenswrapper[4948]: I0220 09:16:08.026333 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:16:20 crc kubenswrapper[4948]: I0220 09:16:20.992121 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hrvz7"] Feb 20 09:16:20 crc kubenswrapper[4948]: E0220 09:16:20.993332 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1443dd18-dcab-42e8-8592-63e1ae18a6ef" containerName="registry-server" Feb 20 09:16:20 crc kubenswrapper[4948]: I0220 09:16:20.993350 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="1443dd18-dcab-42e8-8592-63e1ae18a6ef" containerName="registry-server" Feb 20 09:16:20 crc kubenswrapper[4948]: E0220 09:16:20.993369 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4425c869-8c06-4583-bd41-9fbb012a893e" containerName="registry-server" Feb 20 09:16:20 crc kubenswrapper[4948]: I0220 09:16:20.993378 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="4425c869-8c06-4583-bd41-9fbb012a893e" containerName="registry-server" Feb 20 09:16:20 crc kubenswrapper[4948]: E0220 09:16:20.993398 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1443dd18-dcab-42e8-8592-63e1ae18a6ef" containerName="extract-content" Feb 20 09:16:20 crc kubenswrapper[4948]: I0220 09:16:20.993408 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="1443dd18-dcab-42e8-8592-63e1ae18a6ef" containerName="extract-content" Feb 20 09:16:20 crc kubenswrapper[4948]: E0220 09:16:20.993431 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1443dd18-dcab-42e8-8592-63e1ae18a6ef" containerName="extract-utilities" Feb 20 09:16:20 crc kubenswrapper[4948]: I0220 09:16:20.993439 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="1443dd18-dcab-42e8-8592-63e1ae18a6ef" containerName="extract-utilities" Feb 20 09:16:20 crc kubenswrapper[4948]: E0220 09:16:20.993458 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4425c869-8c06-4583-bd41-9fbb012a893e" containerName="extract-content" Feb 20 09:16:20 crc kubenswrapper[4948]: I0220 09:16:20.993466 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="4425c869-8c06-4583-bd41-9fbb012a893e" containerName="extract-content" Feb 20 09:16:20 crc kubenswrapper[4948]: E0220 09:16:20.993485 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4425c869-8c06-4583-bd41-9fbb012a893e" containerName="extract-utilities" Feb 20 09:16:20 crc kubenswrapper[4948]: I0220 09:16:20.993494 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="4425c869-8c06-4583-bd41-9fbb012a893e" containerName="extract-utilities" Feb 20 09:16:20 crc kubenswrapper[4948]: I0220 09:16:20.993721 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="4425c869-8c06-4583-bd41-9fbb012a893e" containerName="registry-server" Feb 20 09:16:20 crc kubenswrapper[4948]: I0220 09:16:20.993738 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="1443dd18-dcab-42e8-8592-63e1ae18a6ef" containerName="registry-server" Feb 20 09:16:20 crc kubenswrapper[4948]: I0220 09:16:20.998653 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hrvz7" Feb 20 09:16:21 crc kubenswrapper[4948]: I0220 09:16:21.016353 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hrvz7"] Feb 20 09:16:21 crc kubenswrapper[4948]: I0220 09:16:21.160239 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7cqf\" (UniqueName: \"kubernetes.io/projected/69047018-e836-408d-9ff4-f6d9ee4a37b2-kube-api-access-m7cqf\") pod \"redhat-operators-hrvz7\" (UID: \"69047018-e836-408d-9ff4-f6d9ee4a37b2\") " pod="openshift-marketplace/redhat-operators-hrvz7" Feb 20 09:16:21 crc kubenswrapper[4948]: I0220 09:16:21.160354 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69047018-e836-408d-9ff4-f6d9ee4a37b2-catalog-content\") pod \"redhat-operators-hrvz7\" (UID: \"69047018-e836-408d-9ff4-f6d9ee4a37b2\") " pod="openshift-marketplace/redhat-operators-hrvz7" Feb 20 09:16:21 crc kubenswrapper[4948]: I0220 09:16:21.160489 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69047018-e836-408d-9ff4-f6d9ee4a37b2-utilities\") pod \"redhat-operators-hrvz7\" (UID: \"69047018-e836-408d-9ff4-f6d9ee4a37b2\") " pod="openshift-marketplace/redhat-operators-hrvz7" Feb 20 09:16:21 crc kubenswrapper[4948]: I0220 09:16:21.261835 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7cqf\" (UniqueName: \"kubernetes.io/projected/69047018-e836-408d-9ff4-f6d9ee4a37b2-kube-api-access-m7cqf\") pod \"redhat-operators-hrvz7\" (UID: \"69047018-e836-408d-9ff4-f6d9ee4a37b2\") " pod="openshift-marketplace/redhat-operators-hrvz7" Feb 20 09:16:21 crc kubenswrapper[4948]: I0220 09:16:21.262272 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69047018-e836-408d-9ff4-f6d9ee4a37b2-catalog-content\") pod \"redhat-operators-hrvz7\" (UID: \"69047018-e836-408d-9ff4-f6d9ee4a37b2\") " pod="openshift-marketplace/redhat-operators-hrvz7" Feb 20 09:16:21 crc kubenswrapper[4948]: I0220 09:16:21.262444 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69047018-e836-408d-9ff4-f6d9ee4a37b2-utilities\") pod \"redhat-operators-hrvz7\" (UID: \"69047018-e836-408d-9ff4-f6d9ee4a37b2\") " pod="openshift-marketplace/redhat-operators-hrvz7" Feb 20 09:16:21 crc kubenswrapper[4948]: I0220 09:16:21.262899 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69047018-e836-408d-9ff4-f6d9ee4a37b2-catalog-content\") pod \"redhat-operators-hrvz7\" (UID: \"69047018-e836-408d-9ff4-f6d9ee4a37b2\") " pod="openshift-marketplace/redhat-operators-hrvz7" Feb 20 09:16:21 crc kubenswrapper[4948]: I0220 09:16:21.263054 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69047018-e836-408d-9ff4-f6d9ee4a37b2-utilities\") pod \"redhat-operators-hrvz7\" (UID: \"69047018-e836-408d-9ff4-f6d9ee4a37b2\") " pod="openshift-marketplace/redhat-operators-hrvz7" Feb 20 09:16:21 crc kubenswrapper[4948]: I0220 09:16:21.289797 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7cqf\" (UniqueName: \"kubernetes.io/projected/69047018-e836-408d-9ff4-f6d9ee4a37b2-kube-api-access-m7cqf\") pod \"redhat-operators-hrvz7\" (UID: \"69047018-e836-408d-9ff4-f6d9ee4a37b2\") " pod="openshift-marketplace/redhat-operators-hrvz7" Feb 20 09:16:21 crc kubenswrapper[4948]: I0220 09:16:21.347049 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hrvz7" Feb 20 09:16:22 crc kubenswrapper[4948]: I0220 09:16:21.822914 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hrvz7"] Feb 20 09:16:22 crc kubenswrapper[4948]: I0220 09:16:22.914785 4948 generic.go:334] "Generic (PLEG): container finished" podID="69047018-e836-408d-9ff4-f6d9ee4a37b2" containerID="3a5d3c0f1a659e4e232e276ac70ee5562e34ae0a4e76e4f6f29a8672c1834438" exitCode=0 Feb 20 09:16:22 crc kubenswrapper[4948]: I0220 09:16:22.914859 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrvz7" event={"ID":"69047018-e836-408d-9ff4-f6d9ee4a37b2","Type":"ContainerDied","Data":"3a5d3c0f1a659e4e232e276ac70ee5562e34ae0a4e76e4f6f29a8672c1834438"} Feb 20 09:16:22 crc kubenswrapper[4948]: I0220 09:16:22.915251 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrvz7" event={"ID":"69047018-e836-408d-9ff4-f6d9ee4a37b2","Type":"ContainerStarted","Data":"f45caaea490893f7fa52af61c5e11466b24ca3554bf9aff6751d648751c80151"} Feb 20 09:16:23 crc kubenswrapper[4948]: I0220 09:16:23.926673 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrvz7" event={"ID":"69047018-e836-408d-9ff4-f6d9ee4a37b2","Type":"ContainerStarted","Data":"69f6a62bc42ec81d9c5bc2e4e325ac62148fca78690de6243134a1c2fa0d1758"} Feb 20 09:16:26 crc kubenswrapper[4948]: I0220 09:16:26.959235 4948 generic.go:334] "Generic (PLEG): container finished" podID="69047018-e836-408d-9ff4-f6d9ee4a37b2" containerID="69f6a62bc42ec81d9c5bc2e4e325ac62148fca78690de6243134a1c2fa0d1758" exitCode=0 Feb 20 09:16:26 crc kubenswrapper[4948]: I0220 09:16:26.959286 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrvz7" event={"ID":"69047018-e836-408d-9ff4-f6d9ee4a37b2","Type":"ContainerDied","Data":"69f6a62bc42ec81d9c5bc2e4e325ac62148fca78690de6243134a1c2fa0d1758"} Feb 20 09:16:27 crc kubenswrapper[4948]: I0220 09:16:27.970744 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrvz7" event={"ID":"69047018-e836-408d-9ff4-f6d9ee4a37b2","Type":"ContainerStarted","Data":"18ea1ab3b95b4d88bfcc704c5faf23f222ba4425d7abde000ef461eb010ddd32"} Feb 20 09:16:27 crc kubenswrapper[4948]: I0220 09:16:27.996637 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hrvz7" podStartSLOduration=3.569465246 podStartE2EDuration="7.996611869s" podCreationTimestamp="2026-02-20 09:16:20 +0000 UTC" firstStartedPulling="2026-02-20 09:16:22.9168677 +0000 UTC m=+4231.891362530" lastFinishedPulling="2026-02-20 09:16:27.344014303 +0000 UTC m=+4236.318509153" observedRunningTime="2026-02-20 09:16:27.99260372 +0000 UTC m=+4236.967098580" watchObservedRunningTime="2026-02-20 09:16:27.996611869 +0000 UTC m=+4236.971106699" Feb 20 09:16:31 crc kubenswrapper[4948]: I0220 09:16:31.347789 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hrvz7" Feb 20 09:16:31 crc kubenswrapper[4948]: I0220 09:16:31.348521 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hrvz7" Feb 20 09:16:32 crc kubenswrapper[4948]: I0220 09:16:32.430174 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hrvz7" podUID="69047018-e836-408d-9ff4-f6d9ee4a37b2" containerName="registry-server" probeResult="failure" output=< Feb 20 09:16:32 crc kubenswrapper[4948]: timeout: failed to connect service ":50051" within 1s Feb 20 09:16:32 crc kubenswrapper[4948]: > Feb 20 09:16:38 crc kubenswrapper[4948]: I0220 09:16:38.025306 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:16:38 crc kubenswrapper[4948]: I0220 09:16:38.025926 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:16:38 crc kubenswrapper[4948]: I0220 09:16:38.026007 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 09:16:38 crc kubenswrapper[4948]: I0220 09:16:38.026790 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2e609219eda64b1acca0e9ec84aa03d7c21986cd69e1f5ce367d12346cb500e5"} pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 09:16:38 crc kubenswrapper[4948]: I0220 09:16:38.026851 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" containerID="cri-o://2e609219eda64b1acca0e9ec84aa03d7c21986cd69e1f5ce367d12346cb500e5" gracePeriod=600 Feb 20 09:16:38 crc kubenswrapper[4948]: E0220 09:16:38.154363 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:16:39 crc kubenswrapper[4948]: I0220 09:16:39.093727 4948 generic.go:334] "Generic (PLEG): container finished" podID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerID="2e609219eda64b1acca0e9ec84aa03d7c21986cd69e1f5ce367d12346cb500e5" exitCode=0 Feb 20 09:16:39 crc kubenswrapper[4948]: I0220 09:16:39.093769 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerDied","Data":"2e609219eda64b1acca0e9ec84aa03d7c21986cd69e1f5ce367d12346cb500e5"} Feb 20 09:16:39 crc kubenswrapper[4948]: I0220 09:16:39.093800 4948 scope.go:117] "RemoveContainer" containerID="5bbfc48505968074b16a56aea08c035f5c0687419ea24b66d69a44ebd2ab9d7d" Feb 20 09:16:39 crc kubenswrapper[4948]: I0220 09:16:39.094731 4948 scope.go:117] "RemoveContainer" containerID="2e609219eda64b1acca0e9ec84aa03d7c21986cd69e1f5ce367d12346cb500e5" Feb 20 09:16:39 crc kubenswrapper[4948]: E0220 09:16:39.095355 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:16:41 crc kubenswrapper[4948]: I0220 09:16:41.415158 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hrvz7" Feb 20 09:16:41 crc kubenswrapper[4948]: I0220 09:16:41.471857 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hrvz7" Feb 20 09:16:41 crc kubenswrapper[4948]: I0220 09:16:41.656562 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hrvz7"] Feb 20 09:16:43 crc kubenswrapper[4948]: I0220 09:16:43.131219 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hrvz7" podUID="69047018-e836-408d-9ff4-f6d9ee4a37b2" containerName="registry-server" containerID="cri-o://18ea1ab3b95b4d88bfcc704c5faf23f222ba4425d7abde000ef461eb010ddd32" gracePeriod=2 Feb 20 09:16:43 crc kubenswrapper[4948]: I0220 09:16:43.661330 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hrvz7" Feb 20 09:16:43 crc kubenswrapper[4948]: I0220 09:16:43.755153 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69047018-e836-408d-9ff4-f6d9ee4a37b2-catalog-content\") pod \"69047018-e836-408d-9ff4-f6d9ee4a37b2\" (UID: \"69047018-e836-408d-9ff4-f6d9ee4a37b2\") " Feb 20 09:16:43 crc kubenswrapper[4948]: I0220 09:16:43.755236 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7cqf\" (UniqueName: \"kubernetes.io/projected/69047018-e836-408d-9ff4-f6d9ee4a37b2-kube-api-access-m7cqf\") pod \"69047018-e836-408d-9ff4-f6d9ee4a37b2\" (UID: \"69047018-e836-408d-9ff4-f6d9ee4a37b2\") " Feb 20 09:16:43 crc kubenswrapper[4948]: I0220 09:16:43.755520 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69047018-e836-408d-9ff4-f6d9ee4a37b2-utilities\") pod \"69047018-e836-408d-9ff4-f6d9ee4a37b2\" (UID: \"69047018-e836-408d-9ff4-f6d9ee4a37b2\") " Feb 20 09:16:43 crc kubenswrapper[4948]: I0220 09:16:43.756815 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69047018-e836-408d-9ff4-f6d9ee4a37b2-utilities" (OuterVolumeSpecName: "utilities") pod "69047018-e836-408d-9ff4-f6d9ee4a37b2" (UID: "69047018-e836-408d-9ff4-f6d9ee4a37b2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:16:43 crc kubenswrapper[4948]: I0220 09:16:43.765561 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69047018-e836-408d-9ff4-f6d9ee4a37b2-kube-api-access-m7cqf" (OuterVolumeSpecName: "kube-api-access-m7cqf") pod "69047018-e836-408d-9ff4-f6d9ee4a37b2" (UID: "69047018-e836-408d-9ff4-f6d9ee4a37b2"). InnerVolumeSpecName "kube-api-access-m7cqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:16:43 crc kubenswrapper[4948]: I0220 09:16:43.858465 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7cqf\" (UniqueName: \"kubernetes.io/projected/69047018-e836-408d-9ff4-f6d9ee4a37b2-kube-api-access-m7cqf\") on node \"crc\" DevicePath \"\"" Feb 20 09:16:43 crc kubenswrapper[4948]: I0220 09:16:43.858525 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69047018-e836-408d-9ff4-f6d9ee4a37b2-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 09:16:43 crc kubenswrapper[4948]: I0220 09:16:43.905129 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69047018-e836-408d-9ff4-f6d9ee4a37b2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "69047018-e836-408d-9ff4-f6d9ee4a37b2" (UID: "69047018-e836-408d-9ff4-f6d9ee4a37b2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:16:43 crc kubenswrapper[4948]: I0220 09:16:43.960049 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69047018-e836-408d-9ff4-f6d9ee4a37b2-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 09:16:44 crc kubenswrapper[4948]: I0220 09:16:44.144407 4948 generic.go:334] "Generic (PLEG): container finished" podID="69047018-e836-408d-9ff4-f6d9ee4a37b2" containerID="18ea1ab3b95b4d88bfcc704c5faf23f222ba4425d7abde000ef461eb010ddd32" exitCode=0 Feb 20 09:16:44 crc kubenswrapper[4948]: I0220 09:16:44.144461 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrvz7" event={"ID":"69047018-e836-408d-9ff4-f6d9ee4a37b2","Type":"ContainerDied","Data":"18ea1ab3b95b4d88bfcc704c5faf23f222ba4425d7abde000ef461eb010ddd32"} Feb 20 09:16:44 crc kubenswrapper[4948]: I0220 09:16:44.144499 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hrvz7" Feb 20 09:16:44 crc kubenswrapper[4948]: I0220 09:16:44.144532 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrvz7" event={"ID":"69047018-e836-408d-9ff4-f6d9ee4a37b2","Type":"ContainerDied","Data":"f45caaea490893f7fa52af61c5e11466b24ca3554bf9aff6751d648751c80151"} Feb 20 09:16:44 crc kubenswrapper[4948]: I0220 09:16:44.144553 4948 scope.go:117] "RemoveContainer" containerID="18ea1ab3b95b4d88bfcc704c5faf23f222ba4425d7abde000ef461eb010ddd32" Feb 20 09:16:44 crc kubenswrapper[4948]: I0220 09:16:44.168434 4948 scope.go:117] "RemoveContainer" containerID="69f6a62bc42ec81d9c5bc2e4e325ac62148fca78690de6243134a1c2fa0d1758" Feb 20 09:16:44 crc kubenswrapper[4948]: I0220 09:16:44.210192 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hrvz7"] Feb 20 09:16:44 crc kubenswrapper[4948]: I0220 09:16:44.211396 4948 scope.go:117] "RemoveContainer" containerID="3a5d3c0f1a659e4e232e276ac70ee5562e34ae0a4e76e4f6f29a8672c1834438" Feb 20 09:16:44 crc kubenswrapper[4948]: I0220 09:16:44.230008 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hrvz7"] Feb 20 09:16:44 crc kubenswrapper[4948]: I0220 09:16:44.264341 4948 scope.go:117] "RemoveContainer" containerID="18ea1ab3b95b4d88bfcc704c5faf23f222ba4425d7abde000ef461eb010ddd32" Feb 20 09:16:44 crc kubenswrapper[4948]: E0220 09:16:44.264776 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18ea1ab3b95b4d88bfcc704c5faf23f222ba4425d7abde000ef461eb010ddd32\": container with ID starting with 18ea1ab3b95b4d88bfcc704c5faf23f222ba4425d7abde000ef461eb010ddd32 not found: ID does not exist" containerID="18ea1ab3b95b4d88bfcc704c5faf23f222ba4425d7abde000ef461eb010ddd32" Feb 20 09:16:44 crc kubenswrapper[4948]: I0220 09:16:44.264806 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18ea1ab3b95b4d88bfcc704c5faf23f222ba4425d7abde000ef461eb010ddd32"} err="failed to get container status \"18ea1ab3b95b4d88bfcc704c5faf23f222ba4425d7abde000ef461eb010ddd32\": rpc error: code = NotFound desc = could not find container \"18ea1ab3b95b4d88bfcc704c5faf23f222ba4425d7abde000ef461eb010ddd32\": container with ID starting with 18ea1ab3b95b4d88bfcc704c5faf23f222ba4425d7abde000ef461eb010ddd32 not found: ID does not exist" Feb 20 09:16:44 crc kubenswrapper[4948]: I0220 09:16:44.264831 4948 scope.go:117] "RemoveContainer" containerID="69f6a62bc42ec81d9c5bc2e4e325ac62148fca78690de6243134a1c2fa0d1758" Feb 20 09:16:44 crc kubenswrapper[4948]: E0220 09:16:44.265319 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69f6a62bc42ec81d9c5bc2e4e325ac62148fca78690de6243134a1c2fa0d1758\": container with ID starting with 69f6a62bc42ec81d9c5bc2e4e325ac62148fca78690de6243134a1c2fa0d1758 not found: ID does not exist" containerID="69f6a62bc42ec81d9c5bc2e4e325ac62148fca78690de6243134a1c2fa0d1758" Feb 20 09:16:44 crc kubenswrapper[4948]: I0220 09:16:44.265420 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69f6a62bc42ec81d9c5bc2e4e325ac62148fca78690de6243134a1c2fa0d1758"} err="failed to get container status \"69f6a62bc42ec81d9c5bc2e4e325ac62148fca78690de6243134a1c2fa0d1758\": rpc error: code = NotFound desc = could not find container \"69f6a62bc42ec81d9c5bc2e4e325ac62148fca78690de6243134a1c2fa0d1758\": container with ID starting with 69f6a62bc42ec81d9c5bc2e4e325ac62148fca78690de6243134a1c2fa0d1758 not found: ID does not exist" Feb 20 09:16:44 crc kubenswrapper[4948]: I0220 09:16:44.265495 4948 scope.go:117] "RemoveContainer" containerID="3a5d3c0f1a659e4e232e276ac70ee5562e34ae0a4e76e4f6f29a8672c1834438" Feb 20 09:16:44 crc kubenswrapper[4948]: E0220 09:16:44.265834 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a5d3c0f1a659e4e232e276ac70ee5562e34ae0a4e76e4f6f29a8672c1834438\": container with ID starting with 3a5d3c0f1a659e4e232e276ac70ee5562e34ae0a4e76e4f6f29a8672c1834438 not found: ID does not exist" containerID="3a5d3c0f1a659e4e232e276ac70ee5562e34ae0a4e76e4f6f29a8672c1834438" Feb 20 09:16:44 crc kubenswrapper[4948]: I0220 09:16:44.265916 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a5d3c0f1a659e4e232e276ac70ee5562e34ae0a4e76e4f6f29a8672c1834438"} err="failed to get container status \"3a5d3c0f1a659e4e232e276ac70ee5562e34ae0a4e76e4f6f29a8672c1834438\": rpc error: code = NotFound desc = could not find container \"3a5d3c0f1a659e4e232e276ac70ee5562e34ae0a4e76e4f6f29a8672c1834438\": container with ID starting with 3a5d3c0f1a659e4e232e276ac70ee5562e34ae0a4e76e4f6f29a8672c1834438 not found: ID does not exist" Feb 20 09:16:45 crc kubenswrapper[4948]: I0220 09:16:45.736851 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69047018-e836-408d-9ff4-f6d9ee4a37b2" path="/var/lib/kubelet/pods/69047018-e836-408d-9ff4-f6d9ee4a37b2/volumes" Feb 20 09:16:52 crc kubenswrapper[4948]: I0220 09:16:52.722833 4948 scope.go:117] "RemoveContainer" containerID="2e609219eda64b1acca0e9ec84aa03d7c21986cd69e1f5ce367d12346cb500e5" Feb 20 09:16:52 crc kubenswrapper[4948]: E0220 09:16:52.724000 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:17:04 crc kubenswrapper[4948]: I0220 09:17:04.722680 4948 scope.go:117] "RemoveContainer" containerID="2e609219eda64b1acca0e9ec84aa03d7c21986cd69e1f5ce367d12346cb500e5" Feb 20 09:17:04 crc kubenswrapper[4948]: E0220 09:17:04.723616 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:17:18 crc kubenswrapper[4948]: I0220 09:17:18.723552 4948 scope.go:117] "RemoveContainer" containerID="2e609219eda64b1acca0e9ec84aa03d7c21986cd69e1f5ce367d12346cb500e5" Feb 20 09:17:18 crc kubenswrapper[4948]: E0220 09:17:18.724414 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:17:33 crc kubenswrapper[4948]: I0220 09:17:33.732286 4948 scope.go:117] "RemoveContainer" containerID="2e609219eda64b1acca0e9ec84aa03d7c21986cd69e1f5ce367d12346cb500e5" Feb 20 09:17:33 crc kubenswrapper[4948]: E0220 09:17:33.733346 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee"